Apr 21 14:23:21.506010 ip-10-0-141-61 systemd[1]: kubelet.service: Failed to load environment files: No such file or directory Apr 21 14:23:21.506020 ip-10-0-141-61 systemd[1]: kubelet.service: Failed to run 'start-pre' task: No such file or directory Apr 21 14:23:21.506028 ip-10-0-141-61 systemd[1]: kubelet.service: Failed with result 'resources'. Apr 21 14:23:21.506298 ip-10-0-141-61 systemd[1]: Failed to start Kubernetes Kubelet. Apr 21 14:23:31.633475 ip-10-0-141-61 systemd[1]: kubelet.service: Failed to schedule restart job: Unit crio.service not found. Apr 21 14:23:31.633497 ip-10-0-141-61 systemd[1]: kubelet.service: Failed with result 'resources'. -- Boot 0a39a08217e4400faa1fad66213ac647 -- Apr 21 14:26:06.400917 ip-10-0-141-61 systemd[1]: Starting Kubernetes Kubelet... Apr 21 14:26:06.952857 ip-10-0-141-61 kubenswrapper[2580]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 14:26:06.952857 ip-10-0-141-61 kubenswrapper[2580]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 21 14:26:06.952857 ip-10-0-141-61 kubenswrapper[2580]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 14:26:06.952857 ip-10-0-141-61 kubenswrapper[2580]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 21 14:26:06.952857 ip-10-0-141-61 kubenswrapper[2580]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 14:26:06.957190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.957097 2580 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 21 14:26:06.959475 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959457 2580 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 14:26:06.959475 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959475 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959479 2580 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959482 2580 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959485 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959490 2580 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959495 2580 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959498 2580 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959501 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959504 2580 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959507 2580 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959510 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959512 2580 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959516 2580 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959519 2580 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959521 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959525 2580 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959528 2580 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959530 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959533 2580 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959535 2580 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 14:26:06.959540 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959537 2580 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959540 2580 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959543 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959546 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959549 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959552 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959555 2580 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959557 2580 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959575 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959579 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959581 2580 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959584 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959587 2580 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959590 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959593 2580 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959595 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959598 2580 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959600 2580 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959603 2580 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959605 2580 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 14:26:06.960035 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959608 2580 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959612 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959614 2580 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959617 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959619 2580 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959622 2580 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959624 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959627 2580 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959629 2580 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959632 2580 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959634 2580 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959637 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959639 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959642 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959645 2580 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959648 2580 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959650 2580 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959653 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959655 2580 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959658 2580 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 14:26:06.960523 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959661 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959663 2580 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959666 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959668 2580 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959670 2580 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959674 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959676 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959679 2580 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959681 2580 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959684 2580 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959688 2580 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959691 2580 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959695 2580 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959698 2580 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959700 2580 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959704 2580 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959706 2580 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959708 2580 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959711 2580 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 14:26:06.961024 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959713 2580 feature_gate.go:328] unrecognized feature gate: Example Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959716 2580 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959718 2580 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959721 2580 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959724 2580 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.959727 2580 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960137 2580 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960143 2580 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960146 2580 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960149 2580 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960152 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960154 2580 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960157 2580 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960160 2580 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960162 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960165 2580 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960168 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960170 2580 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960173 2580 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960176 2580 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 14:26:06.961487 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960179 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960181 2580 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960184 2580 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960186 2580 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960189 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960191 2580 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960194 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960198 2580 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960202 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960205 2580 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960208 2580 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960211 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960213 2580 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960216 2580 feature_gate.go:328] unrecognized feature gate: Example Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960219 2580 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960221 2580 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960223 2580 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960226 2580 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960229 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960233 2580 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 14:26:06.962017 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960236 2580 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960239 2580 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960241 2580 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960244 2580 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960246 2580 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960249 2580 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960252 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960254 2580 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960257 2580 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960259 2580 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960262 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960264 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960267 2580 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960270 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960272 2580 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960275 2580 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960277 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960280 2580 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960283 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960285 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 14:26:06.962516 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960288 2580 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960291 2580 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960293 2580 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960296 2580 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960298 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960301 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960303 2580 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960306 2580 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960308 2580 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960311 2580 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960313 2580 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960316 2580 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960319 2580 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960321 2580 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960324 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960327 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960329 2580 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960332 2580 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960334 2580 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960337 2580 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 14:26:06.963023 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960339 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960343 2580 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960347 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960351 2580 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960354 2580 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960356 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960359 2580 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960361 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960364 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960366 2580 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960368 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.960371 2580 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961277 2580 flags.go:64] FLAG: --address="0.0.0.0" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961286 2580 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961294 2580 flags.go:64] FLAG: --anonymous-auth="true" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961298 2580 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961303 2580 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961306 2580 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961310 2580 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961315 2580 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961318 2580 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 21 14:26:06.963517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961321 2580 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961324 2580 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961328 2580 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961331 2580 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961334 2580 flags.go:64] FLAG: --cgroup-root="" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961337 2580 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961340 2580 flags.go:64] FLAG: --client-ca-file="" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961343 2580 flags.go:64] FLAG: --cloud-config="" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961346 2580 flags.go:64] FLAG: --cloud-provider="external" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961349 2580 flags.go:64] FLAG: --cluster-dns="[]" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961353 2580 flags.go:64] FLAG: --cluster-domain="" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961356 2580 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961359 2580 flags.go:64] FLAG: --config-dir="" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961362 2580 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961365 2580 flags.go:64] FLAG: --container-log-max-files="5" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961369 2580 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961373 2580 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961376 2580 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961379 2580 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961382 2580 flags.go:64] FLAG: --contention-profiling="false" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961386 2580 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961389 2580 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961392 2580 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961395 2580 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961399 2580 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 21 14:26:06.964044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961403 2580 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961406 2580 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961409 2580 flags.go:64] FLAG: --enable-load-reader="false" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961412 2580 flags.go:64] FLAG: --enable-server="true" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961414 2580 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961419 2580 flags.go:64] FLAG: --event-burst="100" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961422 2580 flags.go:64] FLAG: --event-qps="50" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961425 2580 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961428 2580 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961432 2580 flags.go:64] FLAG: --eviction-hard="" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961436 2580 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961438 2580 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961441 2580 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961444 2580 flags.go:64] FLAG: --eviction-soft="" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961447 2580 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961450 2580 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961453 2580 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961456 2580 flags.go:64] FLAG: --experimental-mounter-path="" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961459 2580 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961462 2580 flags.go:64] FLAG: --fail-swap-on="true" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961465 2580 flags.go:64] FLAG: --feature-gates="" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961469 2580 flags.go:64] FLAG: --file-check-frequency="20s" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961472 2580 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961476 2580 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961480 2580 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961483 2580 flags.go:64] FLAG: --healthz-port="10248" Apr 21 14:26:06.964740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961486 2580 flags.go:64] FLAG: --help="false" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961489 2580 flags.go:64] FLAG: --hostname-override="ip-10-0-141-61.ec2.internal" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961492 2580 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961495 2580 flags.go:64] FLAG: --http-check-frequency="20s" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961498 2580 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961511 2580 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961516 2580 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961519 2580 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961522 2580 flags.go:64] FLAG: --image-service-endpoint="" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961525 2580 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961528 2580 flags.go:64] FLAG: --kube-api-burst="100" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961531 2580 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961534 2580 flags.go:64] FLAG: --kube-api-qps="50" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961537 2580 flags.go:64] FLAG: --kube-reserved="" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961540 2580 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961543 2580 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961546 2580 flags.go:64] FLAG: --kubelet-cgroups="" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961548 2580 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961551 2580 flags.go:64] FLAG: --lock-file="" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961554 2580 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961557 2580 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961560 2580 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961581 2580 flags.go:64] FLAG: --log-json-split-stream="false" Apr 21 14:26:06.965386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961584 2580 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961587 2580 flags.go:64] FLAG: --log-text-split-stream="false" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961590 2580 flags.go:64] FLAG: --logging-format="text" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961592 2580 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961596 2580 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961599 2580 flags.go:64] FLAG: --manifest-url="" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961602 2580 flags.go:64] FLAG: --manifest-url-header="" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961606 2580 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961610 2580 flags.go:64] FLAG: --max-open-files="1000000" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961614 2580 flags.go:64] FLAG: --max-pods="110" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961617 2580 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961620 2580 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961623 2580 flags.go:64] FLAG: --memory-manager-policy="None" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961626 2580 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961629 2580 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961633 2580 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961636 2580 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961645 2580 flags.go:64] FLAG: --node-status-max-images="50" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961648 2580 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961651 2580 flags.go:64] FLAG: --oom-score-adj="-999" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961654 2580 flags.go:64] FLAG: --pod-cidr="" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961657 2580 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961664 2580 flags.go:64] FLAG: --pod-manifest-path="" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961667 2580 flags.go:64] FLAG: --pod-max-pids="-1" Apr 21 14:26:06.965988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961670 2580 flags.go:64] FLAG: --pods-per-core="0" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961673 2580 flags.go:64] FLAG: --port="10250" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961676 2580 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961679 2580 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-0b3e5d5d2504e3b6c" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961683 2580 flags.go:64] FLAG: --qos-reserved="" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961685 2580 flags.go:64] FLAG: --read-only-port="10255" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961689 2580 flags.go:64] FLAG: --register-node="true" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961691 2580 flags.go:64] FLAG: --register-schedulable="true" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961694 2580 flags.go:64] FLAG: --register-with-taints="" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961699 2580 flags.go:64] FLAG: --registry-burst="10" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961702 2580 flags.go:64] FLAG: --registry-qps="5" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961705 2580 flags.go:64] FLAG: --reserved-cpus="" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961707 2580 flags.go:64] FLAG: --reserved-memory="" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961711 2580 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961714 2580 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961717 2580 flags.go:64] FLAG: --rotate-certificates="false" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961720 2580 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961725 2580 flags.go:64] FLAG: --runonce="false" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961728 2580 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961731 2580 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961734 2580 flags.go:64] FLAG: --seccomp-default="false" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961737 2580 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961739 2580 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961743 2580 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961747 2580 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961750 2580 flags.go:64] FLAG: --storage-driver-password="root" Apr 21 14:26:06.966623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961753 2580 flags.go:64] FLAG: --storage-driver-secure="false" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961756 2580 flags.go:64] FLAG: --storage-driver-table="stats" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961759 2580 flags.go:64] FLAG: --storage-driver-user="root" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961762 2580 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961765 2580 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961768 2580 flags.go:64] FLAG: --system-cgroups="" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961771 2580 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961777 2580 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961780 2580 flags.go:64] FLAG: --tls-cert-file="" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961783 2580 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961787 2580 flags.go:64] FLAG: --tls-min-version="" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961790 2580 flags.go:64] FLAG: --tls-private-key-file="" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961793 2580 flags.go:64] FLAG: --topology-manager-policy="none" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961796 2580 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961798 2580 flags.go:64] FLAG: --topology-manager-scope="container" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961801 2580 flags.go:64] FLAG: --v="2" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961806 2580 flags.go:64] FLAG: --version="false" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961810 2580 flags.go:64] FLAG: --vmodule="" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961814 2580 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.961818 2580 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961915 2580 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961919 2580 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961922 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961925 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 14:26:06.967237 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961929 2580 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961932 2580 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961935 2580 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961938 2580 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961940 2580 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961943 2580 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961946 2580 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961949 2580 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961952 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961954 2580 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961957 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961960 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961963 2580 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961965 2580 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961968 2580 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961971 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961974 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961977 2580 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961980 2580 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 14:26:06.967868 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961983 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961986 2580 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961988 2580 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961991 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961994 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961997 2580 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.961999 2580 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962002 2580 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962004 2580 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962007 2580 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962009 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962012 2580 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962014 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962017 2580 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962025 2580 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962028 2580 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962032 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962035 2580 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962042 2580 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 14:26:06.968353 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962045 2580 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962047 2580 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962050 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962053 2580 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962056 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962058 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962061 2580 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962064 2580 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962067 2580 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962070 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962072 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962075 2580 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962078 2580 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962080 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962083 2580 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962085 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962088 2580 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962091 2580 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962093 2580 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962096 2580 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 14:26:06.969135 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962098 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962101 2580 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962104 2580 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962108 2580 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962111 2580 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962114 2580 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962116 2580 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962120 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962123 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962126 2580 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962128 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962132 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962134 2580 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962138 2580 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962140 2580 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962143 2580 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962145 2580 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962148 2580 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962150 2580 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962153 2580 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 14:26:06.970034 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962156 2580 feature_gate.go:328] unrecognized feature gate: Example Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962158 2580 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962161 2580 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.962164 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.963027 2580 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.970444 2580 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.970467 2580 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970540 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970549 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970554 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970559 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970590 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970595 2580 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970600 2580 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970605 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970609 2580 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 14:26:06.970926 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970614 2580 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970618 2580 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970622 2580 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970626 2580 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970630 2580 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970635 2580 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970639 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970643 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970647 2580 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970651 2580 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970655 2580 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970659 2580 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970663 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970667 2580 feature_gate.go:328] unrecognized feature gate: Example Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970671 2580 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970675 2580 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970679 2580 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970684 2580 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970689 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970693 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 14:26:06.971779 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970697 2580 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970701 2580 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970707 2580 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970711 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970715 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970719 2580 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970723 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970727 2580 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970731 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970736 2580 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970740 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970744 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970748 2580 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970753 2580 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970757 2580 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970760 2580 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970764 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970768 2580 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970773 2580 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 14:26:06.972282 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970777 2580 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970781 2580 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970785 2580 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970792 2580 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970799 2580 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970804 2580 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970809 2580 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970814 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970819 2580 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970823 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970828 2580 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970832 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970837 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970841 2580 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970844 2580 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970849 2580 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970854 2580 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970860 2580 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970864 2580 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 14:26:06.972904 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970869 2580 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970873 2580 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970877 2580 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970881 2580 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970885 2580 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970891 2580 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970897 2580 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970901 2580 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970906 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970910 2580 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970914 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970918 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970922 2580 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970926 2580 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970930 2580 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970934 2580 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970938 2580 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970942 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 14:26:06.973632 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.970946 2580 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.970954 2580 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971118 2580 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971126 2580 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971131 2580 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971135 2580 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971140 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971145 2580 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971149 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971153 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971158 2580 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971163 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971168 2580 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971173 2580 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971180 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 14:26:06.974358 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971184 2580 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971188 2580 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971193 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971197 2580 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971201 2580 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971205 2580 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971209 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971213 2580 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971217 2580 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971221 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971225 2580 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971229 2580 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971234 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971238 2580 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971242 2580 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971246 2580 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971250 2580 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971254 2580 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971258 2580 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971262 2580 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 14:26:06.975011 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971266 2580 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971270 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971274 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971278 2580 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971283 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971287 2580 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971291 2580 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971296 2580 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971300 2580 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971304 2580 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971309 2580 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971315 2580 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971321 2580 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971326 2580 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971331 2580 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971335 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971340 2580 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971344 2580 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971348 2580 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 14:26:06.975554 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971353 2580 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971356 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971360 2580 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971365 2580 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971368 2580 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971373 2580 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971377 2580 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971381 2580 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971385 2580 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971390 2580 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971394 2580 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971397 2580 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971402 2580 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971406 2580 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971411 2580 feature_gate.go:328] unrecognized feature gate: Example Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971415 2580 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971418 2580 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971423 2580 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971427 2580 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971432 2580 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 14:26:06.976194 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971436 2580 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971440 2580 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971444 2580 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971448 2580 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971453 2580 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971458 2580 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971462 2580 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971466 2580 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971470 2580 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971474 2580 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971478 2580 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971483 2580 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971486 2580 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:06.971490 2580 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.971498 2580 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 14:26:06.976708 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.972427 2580 server.go:962] "Client rotation is on, will bootstrap in background" Apr 21 14:26:06.977086 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.975756 2580 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 21 14:26:06.977240 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.977225 2580 server.go:1019] "Starting client certificate rotation" Apr 21 14:26:06.977347 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.977329 2580 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 14:26:06.978357 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:06.978346 2580 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 14:26:07.011242 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.011210 2580 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 14:26:07.018955 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.018927 2580 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 14:26:07.039909 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.039883 2580 log.go:25] "Validated CRI v1 runtime API" Apr 21 14:26:07.042089 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.042068 2580 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 14:26:07.047425 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.047411 2580 log.go:25] "Validated CRI v1 image API" Apr 21 14:26:07.048743 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.048726 2580 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 21 14:26:07.053717 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.053694 2580 fs.go:135] Filesystem UUIDs: map[4425e415-8f32-4479-bf91-0943a60bdb19:/dev/nvme0n1p3 7B77-95E7:/dev/nvme0n1p2 f62d7a15-51b4-4b09-a733-658caa79e578:/dev/nvme0n1p4] Apr 21 14:26:07.053788 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.053717 2580 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 21 14:26:07.059736 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.059621 2580 manager.go:217] Machine: {Timestamp:2026-04-21 14:26:07.057467378 +0000 UTC m=+0.514951877 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3101300 MemoryCapacity:32812163072 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2a1fce0134bc7d6ba5ac91bc4273f1 SystemUUID:ec2a1fce-0134-bc7d-6ba5-ac91bc4273f1 BootID:0a39a082-17e4-400f-aa1f-ad66213ac647 Filesystems:[{Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16406081536 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16406081536 Type:vfs Inodes:4005391 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6562435072 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:c0:23:8e:e3:4d Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:c0:23:8e:e3:4d Speed:0 Mtu:9001} {Name:ovs-system MacAddress:4a:96:27:1d:1d:e5 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:32812163072 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:34603008 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 21 14:26:07.060810 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.060799 2580 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 21 14:26:07.060907 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.060895 2580 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 21 14:26:07.061255 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.061233 2580 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 21 14:26:07.061399 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.061257 2580 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-141-61.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 21 14:26:07.061444 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.061410 2580 topology_manager.go:138] "Creating topology manager with none policy" Apr 21 14:26:07.061444 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.061418 2580 container_manager_linux.go:306] "Creating device plugin manager" Apr 21 14:26:07.061444 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.061431 2580 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 14:26:07.061771 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.061756 2580 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-2kgk5" Apr 21 14:26:07.063674 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.063659 2580 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 14:26:07.065623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.065610 2580 state_mem.go:36] "Initialized new in-memory state store" Apr 21 14:26:07.065749 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.065739 2580 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 21 14:26:07.068799 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.068788 2580 kubelet.go:491] "Attempting to sync node with API server" Apr 21 14:26:07.068835 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.068808 2580 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 21 14:26:07.068835 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.068824 2580 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 21 14:26:07.068835 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.068834 2580 kubelet.go:397] "Adding apiserver pod source" Apr 21 14:26:07.068965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.068843 2580 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 21 14:26:07.069119 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.069102 2580 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-2kgk5" Apr 21 14:26:07.071164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.071148 2580 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 14:26:07.071241 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.071176 2580 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 14:26:07.074319 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.074294 2580 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 21 14:26:07.078160 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.078134 2580 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 21 14:26:07.080059 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080039 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 21 14:26:07.080059 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080061 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080068 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080074 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080079 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080090 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080097 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080103 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080111 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080118 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080127 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 21 14:26:07.080165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.080136 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 21 14:26:07.084575 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.084541 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 21 14:26:07.084645 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.084582 2580 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 21 14:26:07.084691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.084636 2580 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 14:26:07.084937 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.084920 2580 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 14:26:07.088388 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.088370 2580 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 21 14:26:07.088474 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.088417 2580 server.go:1295] "Started kubelet" Apr 21 14:26:07.088581 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.088534 2580 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 21 14:26:07.088678 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.088533 2580 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 21 14:26:07.088678 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.088610 2580 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 21 14:26:07.089465 ip-10-0-141-61 systemd[1]: Started Kubernetes Kubelet. Apr 21 14:26:07.089770 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.089755 2580 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 21 14:26:07.089823 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.089811 2580 server.go:317] "Adding debug handlers to kubelet server" Apr 21 14:26:07.090372 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.090350 2580 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-141-61.ec2.internal" not found Apr 21 14:26:07.095903 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.095878 2580 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 21 14:26:07.096655 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.096634 2580 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 21 14:26:07.097606 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.097587 2580 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 21 14:26:07.097699 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.097588 2580 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 21 14:26:07.097699 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.097623 2580 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 21 14:26:07.097808 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.097756 2580 reconstruct.go:97] "Volume reconstruction finished" Apr 21 14:26:07.097808 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.097767 2580 reconciler.go:26] "Reconciler: start to sync state" Apr 21 14:26:07.097808 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.097776 2580 factory.go:55] Registering systemd factory Apr 21 14:26:07.097808 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.097796 2580 factory.go:223] Registration of the systemd container factory successfully Apr 21 14:26:07.097989 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:07.097909 2580 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-61.ec2.internal\" not found" Apr 21 14:26:07.098065 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.098030 2580 factory.go:153] Registering CRI-O factory Apr 21 14:26:07.098065 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.098043 2580 factory.go:223] Registration of the crio container factory successfully Apr 21 14:26:07.098126 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.098095 2580 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 21 14:26:07.098154 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.098130 2580 factory.go:103] Registering Raw factory Apr 21 14:26:07.098154 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.098150 2580 manager.go:1196] Started watching for new ooms in manager Apr 21 14:26:07.098494 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.098483 2580 manager.go:319] Starting recovery of all containers Apr 21 14:26:07.098959 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.098932 2580 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 14:26:07.100107 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:07.100068 2580 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 21 14:26:07.102061 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:07.102005 2580 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-10-0-141-61.ec2.internal\" not found" node="ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.104378 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.104355 2580 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-141-61.ec2.internal" not found Apr 21 14:26:07.110559 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.110366 2580 manager.go:324] Recovery completed Apr 21 14:26:07.116806 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.116791 2580 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 14:26:07.118765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.118693 2580 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-61.ec2.internal" event="NodeHasSufficientMemory" Apr 21 14:26:07.118765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.118734 2580 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-61.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 14:26:07.118765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.118750 2580 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-61.ec2.internal" event="NodeHasSufficientPID" Apr 21 14:26:07.119321 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.119306 2580 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 21 14:26:07.119397 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.119321 2580 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 21 14:26:07.119397 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.119344 2580 state_mem.go:36] "Initialized new in-memory state store" Apr 21 14:26:07.122029 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.122013 2580 policy_none.go:49] "None policy: Start" Apr 21 14:26:07.122029 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.122030 2580 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 21 14:26:07.122141 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.122040 2580 state_mem.go:35] "Initializing new in-memory state store" Apr 21 14:26:07.160305 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.160284 2580 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-141-61.ec2.internal" not found Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.163182 2580 manager.go:341] "Starting Device Plugin manager" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:07.163217 2580 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.163232 2580 server.go:85] "Starting device plugin registration server" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.163503 2580 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.163515 2580 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.163618 2580 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.163692 2580 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.163700 2580 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:07.164333 2580 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 21 14:26:07.171577 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:07.164363 2580 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-141-61.ec2.internal\" not found" Apr 21 14:26:07.239202 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.239099 2580 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 21 14:26:07.240481 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.240466 2580 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 21 14:26:07.240543 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.240499 2580 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 21 14:26:07.240543 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.240520 2580 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 21 14:26:07.240543 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.240526 2580 kubelet.go:2451] "Starting kubelet main sync loop" Apr 21 14:26:07.240704 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:07.240558 2580 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 21 14:26:07.245037 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.245020 2580 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 14:26:07.263802 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.263781 2580 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 14:26:07.264755 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.264737 2580 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-61.ec2.internal" event="NodeHasSufficientMemory" Apr 21 14:26:07.264855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.264773 2580 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-61.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 14:26:07.264855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.264789 2580 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-61.ec2.internal" event="NodeHasSufficientPID" Apr 21 14:26:07.264855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.264820 2580 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.274062 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.274040 2580 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.340877 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.340839 2580 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal"] Apr 21 14:26:07.343233 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.343213 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.343320 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.343214 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.369265 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.369237 2580 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.372690 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.372674 2580 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.385556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.385536 2580 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 14:26:07.492015 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.491933 2580 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 14:26:07.499229 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.499200 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/30f945062f07b4b1388971557d0103c9-config\") pod \"kube-apiserver-proxy-ip-10-0-141-61.ec2.internal\" (UID: \"30f945062f07b4b1388971557d0103c9\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.499309 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.499235 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/f283d499d69cf4025cd4d0f1d7e7bb17-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal\" (UID: \"f283d499d69cf4025cd4d0f1d7e7bb17\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.499309 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.499253 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f283d499d69cf4025cd4d0f1d7e7bb17-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal\" (UID: \"f283d499d69cf4025cd4d0f1d7e7bb17\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.600407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.600367 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/30f945062f07b4b1388971557d0103c9-config\") pod \"kube-apiserver-proxy-ip-10-0-141-61.ec2.internal\" (UID: \"30f945062f07b4b1388971557d0103c9\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.600407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.600401 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/30f945062f07b4b1388971557d0103c9-config\") pod \"kube-apiserver-proxy-ip-10-0-141-61.ec2.internal\" (UID: \"30f945062f07b4b1388971557d0103c9\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.600407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.600414 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/f283d499d69cf4025cd4d0f1d7e7bb17-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal\" (UID: \"f283d499d69cf4025cd4d0f1d7e7bb17\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.600667 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.600431 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f283d499d69cf4025cd4d0f1d7e7bb17-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal\" (UID: \"f283d499d69cf4025cd4d0f1d7e7bb17\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.600667 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.600460 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/f283d499d69cf4025cd4d0f1d7e7bb17-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal\" (UID: \"f283d499d69cf4025cd4d0f1d7e7bb17\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.600667 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.600468 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f283d499d69cf4025cd4d0f1d7e7bb17-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal\" (UID: \"f283d499d69cf4025cd4d0f1d7e7bb17\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.688498 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.688451 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.794649 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.794537 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" Apr 21 14:26:07.976580 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.976529 2580 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 21 14:26:07.977076 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.976700 2580 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 14:26:07.977076 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.976734 2580 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 14:26:07.977076 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:07.976741 2580 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 14:26:08.069473 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.069389 2580 apiserver.go:52] "Watching apiserver" Apr 21 14:26:08.072343 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.072310 2580 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-20 14:21:07 +0000 UTC" deadline="2027-11-01 09:40:44.234738622 +0000 UTC" Apr 21 14:26:08.072398 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.072344 2580 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="13411h14m36.162397167s" Apr 21 14:26:08.081606 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.080711 2580 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 21 14:26:08.081755 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.081653 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lmrrn","openshift-image-registry/node-ca-89ch8","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal","openshift-network-diagnostics/network-check-target-j6wgt","openshift-network-operator/iptables-alerter-rjjmt","openshift-cluster-node-tuning-operator/tuned-ld9pw","openshift-multus/multus-8rftw","openshift-multus/multus-additional-cni-plugins-sz9sg","openshift-multus/network-metrics-daemon-fl65d","openshift-ovn-kubernetes/ovnkube-node-ts88z","kube-system/konnectivity-agent-jzct4","kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d"] Apr 21 14:26:08.085110 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.085084 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.086129 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.086107 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.087200 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.087175 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:08.087307 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.087207 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-gcqgm\"" Apr 21 14:26:08.087307 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.087207 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 21 14:26:08.087307 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.087269 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.087307 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.087256 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:08.087586 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.087553 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 21 14:26:08.088358 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.088338 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.088751 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.088737 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 21 14:26:08.088832 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.088817 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-t6cq2\"" Apr 21 14:26:08.089179 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.089167 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 21 14:26:08.089548 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.089532 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.089745 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.089716 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 21 14:26:08.090000 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.089985 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 21 14:26:08.090621 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.090606 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 21 14:26:08.090721 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.090661 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.090912 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.090891 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-vn6nz\"" Apr 21 14:26:08.090990 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.090920 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-fq8sr\"" Apr 21 14:26:08.091095 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091080 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 21 14:26:08.091147 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091119 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 21 14:26:08.091202 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091169 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 21 14:26:08.091412 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091397 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 21 14:26:08.091626 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091613 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 21 14:26:08.091747 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091731 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 21 14:26:08.091927 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091913 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:08.091985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.091972 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-bg7fp\"" Apr 21 14:26:08.091985 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.091974 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:08.092157 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.092142 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 21 14:26:08.092757 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.092732 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 21 14:26:08.092866 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.092787 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 21 14:26:08.092866 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.092844 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-z4mpj\"" Apr 21 14:26:08.093324 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.093308 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.094428 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.094410 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.095158 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095141 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 21 14:26:08.095260 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095173 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-66n5s\"" Apr 21 14:26:08.095260 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095174 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 21 14:26:08.095400 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095291 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 21 14:26:08.095463 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095412 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 21 14:26:08.095463 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095414 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 21 14:26:08.095785 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095769 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.095849 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095832 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 21 14:26:08.095980 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.095962 2580 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 21 14:26:08.096476 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.096463 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-trsck\"" Apr 21 14:26:08.097482 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.097467 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 21 14:26:08.098050 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.098033 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 21 14:26:08.098577 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.098538 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 21 14:26:08.098821 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.098807 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 21 14:26:08.098901 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.098850 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-z72t5\"" Apr 21 14:26:08.098901 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.098884 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 21 14:26:08.099025 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.099006 2580 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 21 14:26:08.104088 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104059 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-host-slash\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.104412 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104398 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-run\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.104470 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104422 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-os-release\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.104470 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104442 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl7wp\" (UniqueName: \"kubernetes.io/projected/ef45360a-730f-4594-a33e-4ae5f5cd4757-kube-api-access-bl7wp\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.104470 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104465 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-cnibin\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.104632 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104494 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-run-ovn-kubernetes\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.104632 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104550 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-etc-selinux\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.104724 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104631 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccae8641-954f-48b3-85eb-812164adcec6-tmp-dir\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.104724 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104677 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-cni-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.104724 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104714 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-hostroot\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.104875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104739 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.104875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104773 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ccae8641-954f-48b3-85eb-812164adcec6-hosts-file\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.104875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104798 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.104875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104833 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-registration-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.104875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104862 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-iptables-alerter-script\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104893 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-kubernetes\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104918 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpc4p\" (UniqueName: \"kubernetes.io/projected/29a15288-a939-4c42-b167-118380ab5462-kube-api-access-rpc4p\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104946 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9ed791bd-825b-4608-ad66-0304c5a34a75-serviceca\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104970 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-sys\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.104994 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-lib-modules\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105016 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-var-lib-kubelet\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105064 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-system-cni-dir\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.105105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105090 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-systemd-units\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105118 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-ovn\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105142 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvrbf\" (UniqueName: \"kubernetes.io/projected/05386859-dd8b-49f4-8271-44949e102c17-kube-api-access-wvrbf\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105166 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/17883801-bbff-4ed7-8ba5-44081b8c5476-agent-certs\") pod \"konnectivity-agent-jzct4\" (UID: \"17883801-bbff-4ed7-8ba5-44081b8c5476\") " pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105204 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-socket-dir-parent\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105229 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105252 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgrd5\" (UniqueName: \"kubernetes.io/projected/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-kube-api-access-dgrd5\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105283 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-cni-netd\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105306 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/05386859-dd8b-49f4-8271-44949e102c17-ovn-node-metrics-cert\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105330 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-sys-fs\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105350 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-host\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105384 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-etc-kubernetes\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105407 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-socket-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105430 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-device-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105458 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-k8s-cni-cncf-io\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.105477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105481 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-daemon-config\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105508 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-multus-certs\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105540 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-ovnkube-script-lib\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105590 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-modprobe-d\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105610 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysconfig\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105626 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105665 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-node-log\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105687 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-system-cni-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105700 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-cnibin\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105714 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef45360a-730f-4594-a33e-4ae5f5cd4757-cni-binary-copy\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105729 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-cni-bin\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105746 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-log-socket\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105765 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-env-overrides\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105778 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-systemd\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105791 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-netns\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105806 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-systemd\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105820 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-etc-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105834 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-kubelet-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105850 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gj98\" (UniqueName: \"kubernetes.io/projected/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-kube-api-access-8gj98\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105864 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysctl-d\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105890 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-cni-multus\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105913 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105949 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.105980 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-ovnkube-config\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106006 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfwqq\" (UniqueName: \"kubernetes.io/projected/ccae8641-954f-48b3-85eb-812164adcec6-kube-api-access-nfwqq\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106032 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-tuned\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106059 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/77abbe16-9a16-46f3-9e44-0d4c056849cd-tmp\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106087 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k55b\" (UniqueName: \"kubernetes.io/projected/77abbe16-9a16-46f3-9e44-0d4c056849cd-kube-api-access-8k55b\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106119 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-cni-binary-copy\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106143 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-slash\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106166 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-cni-bin\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106189 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/17883801-bbff-4ed7-8ba5-44081b8c5476-konnectivity-ca\") pod \"konnectivity-agent-jzct4\" (UID: \"17883801-bbff-4ed7-8ba5-44081b8c5476\") " pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106220 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-kubelet\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.106705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106249 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-os-release\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106271 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106294 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6z2w\" (UniqueName: \"kubernetes.io/projected/9ed791bd-825b-4608-ad66-0304c5a34a75-kube-api-access-k6z2w\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106317 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-kubelet\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106340 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-run-netns\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106362 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdbm9\" (UniqueName: \"kubernetes.io/projected/a4e7a81b-2051-4a16-947b-2430ad7a69ec-kube-api-access-qdbm9\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106385 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysctl-conf\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106409 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-conf-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106430 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ed791bd-825b-4608-ad66-0304c5a34a75-host\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106454 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-var-lib-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.107190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.106966 2580 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 14:26:08.132928 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.132906 2580 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-jp7x6" Apr 21 14:26:08.141997 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.141978 2580 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-jp7x6" Apr 21 14:26:08.207044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207006 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysctl-conf\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.207044 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207038 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-conf-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207057 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ed791bd-825b-4608-ad66-0304c5a34a75-host\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207082 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-var-lib-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207103 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-host-slash\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207140 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-conf-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207160 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9ed791bd-825b-4608-ad66-0304c5a34a75-host\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207147 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-var-lib-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207182 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysctl-conf\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207205 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-run\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207242 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-os-release\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207245 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-host-slash\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207271 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bl7wp\" (UniqueName: \"kubernetes.io/projected/ef45360a-730f-4594-a33e-4ae5f5cd4757-kube-api-access-bl7wp\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207299 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-cnibin\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.207317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207311 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-os-release\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207328 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-run-ovn-kubernetes\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207302 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-run\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207350 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-cnibin\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207374 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-run-ovn-kubernetes\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207414 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-etc-selinux\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207442 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccae8641-954f-48b3-85eb-812164adcec6-tmp-dir\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207466 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-cni-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207504 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-hostroot\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207512 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-etc-selinux\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207522 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207538 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ccae8641-954f-48b3-85eb-812164adcec6-hosts-file\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207555 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-cni-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207582 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207609 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-registration-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207617 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ccae8641-954f-48b3-85eb-812164adcec6-hosts-file\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207615 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-hostroot\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.207876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207659 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207662 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207693 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-iptables-alerter-script\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207718 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-kubernetes\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207741 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rpc4p\" (UniqueName: \"kubernetes.io/projected/29a15288-a939-4c42-b167-118380ab5462-kube-api-access-rpc4p\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207761 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9ed791bd-825b-4608-ad66-0304c5a34a75-serviceca\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207794 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-sys\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207797 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccae8641-954f-48b3-85eb-812164adcec6-tmp-dir\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207815 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-lib-modules\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207842 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-registration-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207841 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-var-lib-kubelet\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207799 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-kubernetes\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207914 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-var-lib-kubelet\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207936 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-lib-modules\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207948 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-system-cni-dir\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207976 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-sys\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.207998 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-systemd-units\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.208668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208014 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-system-cni-dir\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208030 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-ovn\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208058 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wvrbf\" (UniqueName: \"kubernetes.io/projected/05386859-dd8b-49f4-8271-44949e102c17-kube-api-access-wvrbf\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208060 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-systemd-units\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208084 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/17883801-bbff-4ed7-8ba5-44081b8c5476-agent-certs\") pod \"konnectivity-agent-jzct4\" (UID: \"17883801-bbff-4ed7-8ba5-44081b8c5476\") " pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208123 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-ovn\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208122 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-socket-dir-parent\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208165 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-iptables-alerter-script\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208174 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-socket-dir-parent\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208339 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9ed791bd-825b-4608-ad66-0304c5a34a75-serviceca\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208557 2580 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208171 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208701 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dgrd5\" (UniqueName: \"kubernetes.io/projected/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-kube-api-access-dgrd5\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208727 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-cni-netd\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208752 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/05386859-dd8b-49f4-8271-44949e102c17-ovn-node-metrics-cert\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208774 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-sys-fs\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208800 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-host\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.209341 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208799 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208822 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-etc-kubernetes\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208849 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-socket-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208871 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-device-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208878 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-sys-fs\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208893 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-k8s-cni-cncf-io\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208901 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-cni-netd\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208917 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-daemon-config\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.208975 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-multus-certs\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209004 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-ovnkube-script-lib\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209027 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-modprobe-d\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209047 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysconfig\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209049 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-socket-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209070 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209092 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-node-log\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209115 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-system-cni-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209646 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-etc-kubernetes\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.209717 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-multus-certs\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210215 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-cnibin\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210258 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef45360a-730f-4594-a33e-4ae5f5cd4757-cni-binary-copy\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210297 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-cni-bin\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210333 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-log-socket\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210369 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-env-overrides\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210402 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-systemd\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210437 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-netns\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210469 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-systemd\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210527 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-etc-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210579 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-kubelet-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210617 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8gj98\" (UniqueName: \"kubernetes.io/projected/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-kube-api-access-8gj98\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210644 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysctl-d\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210677 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-cni-multus\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210710 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.210772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210744 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210780 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-ovnkube-config\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210811 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfwqq\" (UniqueName: \"kubernetes.io/projected/ccae8641-954f-48b3-85eb-812164adcec6-kube-api-access-nfwqq\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210847 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-tuned\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210883 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/77abbe16-9a16-46f3-9e44-0d4c056849cd-tmp\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210917 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8k55b\" (UniqueName: \"kubernetes.io/projected/77abbe16-9a16-46f3-9e44-0d4c056849cd-kube-api-access-8k55b\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210952 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-cni-binary-copy\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.210986 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-slash\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211020 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-cni-bin\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211054 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/17883801-bbff-4ed7-8ba5-44081b8c5476-konnectivity-ca\") pod \"konnectivity-agent-jzct4\" (UID: \"17883801-bbff-4ed7-8ba5-44081b8c5476\") " pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211080 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-kubelet\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211110 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-os-release\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211167 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211200 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-k6z2w\" (UniqueName: \"kubernetes.io/projected/9ed791bd-825b-4608-ad66-0304c5a34a75-kube-api-access-k6z2w\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211236 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-kubelet\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211264 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-run-netns\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211303 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qdbm9\" (UniqueName: \"kubernetes.io/projected/a4e7a81b-2051-4a16-947b-2430ad7a69ec-kube-api-access-qdbm9\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.211377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211332 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-ovnkube-script-lib\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.212101 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211901 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef45360a-730f-4594-a33e-4ae5f5cd4757-cni-binary-copy\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.212101 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.211977 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-cni-bin\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.212101 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212015 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-log-socket\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.212286 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212266 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-env-overrides\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.212423 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212345 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-systemd\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.212423 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212406 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-modprobe-d\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.214436 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212467 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysconfig\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212677 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-sysctl-d\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212725 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-netns\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212768 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-run-systemd\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214486 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/77abbe16-9a16-46f3-9e44-0d4c056849cd-host\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212860 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-kubelet-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213057 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-cni-binary-copy\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213107 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-slash\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213148 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-cni-bin\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213241 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/17883801-bbff-4ed7-8ba5-44081b8c5476-agent-certs\") pod \"konnectivity-agent-jzct4\" (UID: \"17883801-bbff-4ed7-8ba5-44081b8c5476\") " pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213305 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/a4e7a81b-2051-4a16-947b-2430ad7a69ec-device-dir\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213313 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/05386859-dd8b-49f4-8271-44949e102c17-ovn-node-metrics-cert\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213349 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-run-k8s-cni-cncf-io\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213637 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/17883801-bbff-4ed7-8ba5-44081b8c5476-konnectivity-ca\") pod \"konnectivity-agent-jzct4\" (UID: \"17883801-bbff-4ed7-8ba5-44081b8c5476\") " pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214649 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ef45360a-730f-4594-a33e-4ae5f5cd4757-multus-daemon-config\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213751 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-os-release\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213787 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/05386859-dd8b-49f4-8271-44949e102c17-ovnkube-config\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213832 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-cni-multus\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.213852 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.214801 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:26:08.714772445 +0000 UTC m=+2.172256947 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214123 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-kubelet\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214167 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-host-run-netns\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214343 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/29a15288-a939-4c42-b167-118380ab5462-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214355 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-system-cni-dir\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214377 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-node-log\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.214965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.214401 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-cnibin\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.215703 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.212808 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/05386859-dd8b-49f4-8271-44949e102c17-etc-openvswitch\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.215703 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213692 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef45360a-730f-4594-a33e-4ae5f5cd4757-host-var-lib-kubelet\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.215703 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.213936 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/29a15288-a939-4c42-b167-118380ab5462-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.215971 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.215933 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/77abbe16-9a16-46f3-9e44-0d4c056849cd-tmp\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.217051 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.217030 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/77abbe16-9a16-46f3-9e44-0d4c056849cd-etc-tuned\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.234610 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.234554 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgrd5\" (UniqueName: \"kubernetes.io/projected/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-kube-api-access-dgrd5\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:08.234727 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.234628 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvrbf\" (UniqueName: \"kubernetes.io/projected/05386859-dd8b-49f4-8271-44949e102c17-kube-api-access-wvrbf\") pod \"ovnkube-node-ts88z\" (UID: \"05386859-dd8b-49f4-8271-44949e102c17\") " pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.235522 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.235502 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 14:26:08.235644 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.235528 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 14:26:08.235644 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.235542 2580 projected.go:194] Error preparing data for projected volume kube-api-access-nfdjb for pod openshift-network-diagnostics/network-check-target-j6wgt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:08.235644 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.235626 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb podName:5578d992-9c91-4c3b-8499-59e3e5261890 nodeName:}" failed. No retries permitted until 2026-04-21 14:26:08.735606331 +0000 UTC m=+2.193090836 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nfdjb" (UniqueName: "kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb") pod "network-check-target-j6wgt" (UID: "5578d992-9c91-4c3b-8499-59e3e5261890") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:08.235644 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.235626 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl7wp\" (UniqueName: \"kubernetes.io/projected/ef45360a-730f-4594-a33e-4ae5f5cd4757-kube-api-access-bl7wp\") pod \"multus-8rftw\" (UID: \"ef45360a-730f-4594-a33e-4ae5f5cd4757\") " pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.235858 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.235704 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:08.238127 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.238091 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfwqq\" (UniqueName: \"kubernetes.io/projected/ccae8641-954f-48b3-85eb-812164adcec6-kube-api-access-nfwqq\") pod \"node-resolver-lmrrn\" (UID: \"ccae8641-954f-48b3-85eb-812164adcec6\") " pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.238525 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.238494 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6z2w\" (UniqueName: \"kubernetes.io/projected/9ed791bd-825b-4608-ad66-0304c5a34a75-kube-api-access-k6z2w\") pod \"node-ca-89ch8\" (UID: \"9ed791bd-825b-4608-ad66-0304c5a34a75\") " pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.238768 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.238753 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gj98\" (UniqueName: \"kubernetes.io/projected/45cf2d21-c2f8-43b1-95f7-d3afd9014c36-kube-api-access-8gj98\") pod \"iptables-alerter-rjjmt\" (UID: \"45cf2d21-c2f8-43b1-95f7-d3afd9014c36\") " pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.238999 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.238980 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdbm9\" (UniqueName: \"kubernetes.io/projected/a4e7a81b-2051-4a16-947b-2430ad7a69ec-kube-api-access-qdbm9\") pod \"aws-ebs-csi-driver-node-5g29d\" (UID: \"a4e7a81b-2051-4a16-947b-2430ad7a69ec\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.239477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.239459 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpc4p\" (UniqueName: \"kubernetes.io/projected/29a15288-a939-4c42-b167-118380ab5462-kube-api-access-rpc4p\") pod \"multus-additional-cni-plugins-sz9sg\" (UID: \"29a15288-a939-4c42-b167-118380ab5462\") " pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.241535 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.241514 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k55b\" (UniqueName: \"kubernetes.io/projected/77abbe16-9a16-46f3-9e44-0d4c056849cd-kube-api-access-8k55b\") pod \"tuned-ld9pw\" (UID: \"77abbe16-9a16-46f3-9e44-0d4c056849cd\") " pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.247751 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.247734 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:08.252242 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.252225 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" Apr 21 14:26:08.281324 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.281285 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf283d499d69cf4025cd4d0f1d7e7bb17.slice/crio-d6a302906ca893a6c15ba8b81f89f857c0a2345c93f0e32d43aec326b7b813a3 WatchSource:0}: Error finding container d6a302906ca893a6c15ba8b81f89f857c0a2345c93f0e32d43aec326b7b813a3: Status 404 returned error can't find the container with id d6a302906ca893a6c15ba8b81f89f857c0a2345c93f0e32d43aec326b7b813a3 Apr 21 14:26:08.281818 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.281796 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4e7a81b_2051_4a16_947b_2430ad7a69ec.slice/crio-c2c3507eeca1b80e2a822bd59b21d6a353cf93a5876903ce435f401c537b709d WatchSource:0}: Error finding container c2c3507eeca1b80e2a822bd59b21d6a353cf93a5876903ce435f401c537b709d: Status 404 returned error can't find the container with id c2c3507eeca1b80e2a822bd59b21d6a353cf93a5876903ce435f401c537b709d Apr 21 14:26:08.282383 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.282362 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30f945062f07b4b1388971557d0103c9.slice/crio-2c248310779e82e975fd93f957d1e0a8d228502d23b254ff0f2ae7a2507cc4a8 WatchSource:0}: Error finding container 2c248310779e82e975fd93f957d1e0a8d228502d23b254ff0f2ae7a2507cc4a8: Status 404 returned error can't find the container with id 2c248310779e82e975fd93f957d1e0a8d228502d23b254ff0f2ae7a2507cc4a8 Apr 21 14:26:08.286019 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.286004 2580 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 14:26:08.413875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.413788 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lmrrn" Apr 21 14:26:08.421835 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.421806 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccae8641_954f_48b3_85eb_812164adcec6.slice/crio-4c4ff7a0d332bcc7e57b9e035e8571d9f585cd2440781ee612fbc9d3af8844b4 WatchSource:0}: Error finding container 4c4ff7a0d332bcc7e57b9e035e8571d9f585cd2440781ee612fbc9d3af8844b4: Status 404 returned error can't find the container with id 4c4ff7a0d332bcc7e57b9e035e8571d9f585cd2440781ee612fbc9d3af8844b4 Apr 21 14:26:08.427784 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.427762 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-89ch8" Apr 21 14:26:08.434798 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.434773 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ed791bd_825b_4608_ad66_0304c5a34a75.slice/crio-2916084d45ce5889df91d5fec5df6400740d9e49a3832a82dd0b550ed647243b WatchSource:0}: Error finding container 2916084d45ce5889df91d5fec5df6400740d9e49a3832a82dd0b550ed647243b: Status 404 returned error can't find the container with id 2916084d45ce5889df91d5fec5df6400740d9e49a3832a82dd0b550ed647243b Apr 21 14:26:08.455818 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.455789 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-rjjmt" Apr 21 14:26:08.461800 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.461776 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45cf2d21_c2f8_43b1_95f7_d3afd9014c36.slice/crio-aa1c85c4404b8f24246f1d9d290dd0597bdda712b87b9d3769e20e60bc573f04 WatchSource:0}: Error finding container aa1c85c4404b8f24246f1d9d290dd0597bdda712b87b9d3769e20e60bc573f04: Status 404 returned error can't find the container with id aa1c85c4404b8f24246f1d9d290dd0597bdda712b87b9d3769e20e60bc573f04 Apr 21 14:26:08.472973 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.472949 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" Apr 21 14:26:08.479117 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.479094 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77abbe16_9a16_46f3_9e44_0d4c056849cd.slice/crio-eb6de507d466f0b3ddced8cebca0436a2e9692f16186dc9404f45c5ad773fb3f WatchSource:0}: Error finding container eb6de507d466f0b3ddced8cebca0436a2e9692f16186dc9404f45c5ad773fb3f: Status 404 returned error can't find the container with id eb6de507d466f0b3ddced8cebca0436a2e9692f16186dc9404f45c5ad773fb3f Apr 21 14:26:08.487800 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.487780 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8rftw" Apr 21 14:26:08.493828 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.493796 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef45360a_730f_4594_a33e_4ae5f5cd4757.slice/crio-44f6a01745bdbd5c0816ade70d053df8c8a669d1fb92741c8495fe7b88eb12b3 WatchSource:0}: Error finding container 44f6a01745bdbd5c0816ade70d053df8c8a669d1fb92741c8495fe7b88eb12b3: Status 404 returned error can't find the container with id 44f6a01745bdbd5c0816ade70d053df8c8a669d1fb92741c8495fe7b88eb12b3 Apr 21 14:26:08.508113 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.508094 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" Apr 21 14:26:08.516038 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.516015 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29a15288_a939_4c42_b167_118380ab5462.slice/crio-c2470048ae900880f5bd55aec7ce3f65a9da97d87d401b815064982a2ebd765e WatchSource:0}: Error finding container c2470048ae900880f5bd55aec7ce3f65a9da97d87d401b815064982a2ebd765e: Status 404 returned error can't find the container with id c2470048ae900880f5bd55aec7ce3f65a9da97d87d401b815064982a2ebd765e Apr 21 14:26:08.609170 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.609134 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05386859_dd8b_49f4_8271_44949e102c17.slice/crio-a22838855467773f4d4fbc48f13b945901e7d9d3ebd2740613b55f7f72e31ad5 WatchSource:0}: Error finding container a22838855467773f4d4fbc48f13b945901e7d9d3ebd2740613b55f7f72e31ad5: Status 404 returned error can't find the container with id a22838855467773f4d4fbc48f13b945901e7d9d3ebd2740613b55f7f72e31ad5 Apr 21 14:26:08.632817 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:26:08.632778 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17883801_bbff_4ed7_8ba5_44081b8c5476.slice/crio-ed5a02e40b1a93d2fe3b015a93d549e24693f9356b2fd02dcb8a8d69b48a7e68 WatchSource:0}: Error finding container ed5a02e40b1a93d2fe3b015a93d549e24693f9356b2fd02dcb8a8d69b48a7e68: Status 404 returned error can't find the container with id ed5a02e40b1a93d2fe3b015a93d549e24693f9356b2fd02dcb8a8d69b48a7e68 Apr 21 14:26:08.716583 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.716460 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:08.716737 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.716657 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:08.716737 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.716720 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:26:09.7167018 +0000 UTC m=+3.174186288 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:08.817177 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.817138 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:08.817350 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.817327 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 14:26:08.817428 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.817352 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 14:26:08.817428 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.817374 2580 projected.go:194] Error preparing data for projected volume kube-api-access-nfdjb for pod openshift-network-diagnostics/network-check-target-j6wgt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:08.817558 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:08.817447 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb podName:5578d992-9c91-4c3b-8499-59e3e5261890 nodeName:}" failed. No retries permitted until 2026-04-21 14:26:09.817425925 +0000 UTC m=+3.274910432 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-nfdjb" (UniqueName: "kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb") pod "network-check-target-j6wgt" (UID: "5578d992-9c91-4c3b-8499-59e3e5261890") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:08.944086 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:08.943868 2580 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 14:26:09.102556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.102525 2580 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 14:26:09.132854 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.132822 2580 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 14:26:09.143008 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.142967 2580 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-20 14:21:08 +0000 UTC" deadline="2027-09-18 03:53:08.274115994 +0000 UTC" Apr 21 14:26:09.143008 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.143006 2580 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="12349h26m59.13111371s" Apr 21 14:26:09.251164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.251093 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lmrrn" event={"ID":"ccae8641-954f-48b3-85eb-812164adcec6","Type":"ContainerStarted","Data":"4c4ff7a0d332bcc7e57b9e035e8571d9f585cd2440781ee612fbc9d3af8844b4"} Apr 21 14:26:09.255151 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.255119 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" event={"ID":"a4e7a81b-2051-4a16-947b-2430ad7a69ec","Type":"ContainerStarted","Data":"c2c3507eeca1b80e2a822bd59b21d6a353cf93a5876903ce435f401c537b709d"} Apr 21 14:26:09.261388 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.261350 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"a22838855467773f4d4fbc48f13b945901e7d9d3ebd2740613b55f7f72e31ad5"} Apr 21 14:26:09.287945 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.287905 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerStarted","Data":"c2470048ae900880f5bd55aec7ce3f65a9da97d87d401b815064982a2ebd765e"} Apr 21 14:26:09.293067 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.293032 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-rjjmt" event={"ID":"45cf2d21-c2f8-43b1-95f7-d3afd9014c36","Type":"ContainerStarted","Data":"aa1c85c4404b8f24246f1d9d290dd0597bdda712b87b9d3769e20e60bc573f04"} Apr 21 14:26:09.313417 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.313377 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" event={"ID":"30f945062f07b4b1388971557d0103c9","Type":"ContainerStarted","Data":"2c248310779e82e975fd93f957d1e0a8d228502d23b254ff0f2ae7a2507cc4a8"} Apr 21 14:26:09.315865 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.315830 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" event={"ID":"f283d499d69cf4025cd4d0f1d7e7bb17","Type":"ContainerStarted","Data":"d6a302906ca893a6c15ba8b81f89f857c0a2345c93f0e32d43aec326b7b813a3"} Apr 21 14:26:09.330229 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.330189 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-jzct4" event={"ID":"17883801-bbff-4ed7-8ba5-44081b8c5476","Type":"ContainerStarted","Data":"ed5a02e40b1a93d2fe3b015a93d549e24693f9356b2fd02dcb8a8d69b48a7e68"} Apr 21 14:26:09.349470 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.349377 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rftw" event={"ID":"ef45360a-730f-4594-a33e-4ae5f5cd4757","Type":"ContainerStarted","Data":"44f6a01745bdbd5c0816ade70d053df8c8a669d1fb92741c8495fe7b88eb12b3"} Apr 21 14:26:09.370329 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.370285 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" event={"ID":"77abbe16-9a16-46f3-9e44-0d4c056849cd","Type":"ContainerStarted","Data":"eb6de507d466f0b3ddced8cebca0436a2e9692f16186dc9404f45c5ad773fb3f"} Apr 21 14:26:09.390498 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.390457 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-89ch8" event={"ID":"9ed791bd-825b-4608-ad66-0304c5a34a75","Type":"ContainerStarted","Data":"2916084d45ce5889df91d5fec5df6400740d9e49a3832a82dd0b550ed647243b"} Apr 21 14:26:09.725788 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.725704 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:09.725951 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:09.725813 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:09.725951 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:09.725862 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:26:11.725847988 +0000 UTC m=+5.183332473 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:09.826979 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:09.826946 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:09.827166 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:09.827132 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 14:26:09.827166 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:09.827155 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 14:26:09.827279 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:09.827169 2580 projected.go:194] Error preparing data for projected volume kube-api-access-nfdjb for pod openshift-network-diagnostics/network-check-target-j6wgt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:09.827279 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:09.827224 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb podName:5578d992-9c91-4c3b-8499-59e3e5261890 nodeName:}" failed. No retries permitted until 2026-04-21 14:26:11.82720726 +0000 UTC m=+5.284691760 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-nfdjb" (UniqueName: "kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb") pod "network-check-target-j6wgt" (UID: "5578d992-9c91-4c3b-8499-59e3e5261890") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:10.144470 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:10.143341 2580 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-20 14:21:08 +0000 UTC" deadline="2027-11-01 11:06:55.422618893 +0000 UTC" Apr 21 14:26:10.144470 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:10.143378 2580 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="13412h40m45.279245613s" Apr 21 14:26:10.241663 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:10.241184 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:10.241663 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:10.241292 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:10.241663 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:10.241301 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:10.241663 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:10.241381 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:11.742944 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:11.742847 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:11.743415 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:11.742997 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:11.743415 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:11.743066 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:26:15.743045682 +0000 UTC m=+9.200530171 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:11.843400 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:11.843355 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:11.843582 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:11.843530 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 14:26:11.843582 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:11.843552 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 14:26:11.843582 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:11.843582 2580 projected.go:194] Error preparing data for projected volume kube-api-access-nfdjb for pod openshift-network-diagnostics/network-check-target-j6wgt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:11.843731 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:11.843633 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb podName:5578d992-9c91-4c3b-8499-59e3e5261890 nodeName:}" failed. No retries permitted until 2026-04-21 14:26:15.843618634 +0000 UTC m=+9.301103119 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-nfdjb" (UniqueName: "kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb") pod "network-check-target-j6wgt" (UID: "5578d992-9c91-4c3b-8499-59e3e5261890") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:12.241047 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:12.240957 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:12.241205 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:12.241124 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:12.241556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:12.241527 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:12.241692 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:12.241666 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:14.241193 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:14.241151 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:14.241674 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:14.241301 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:14.241750 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:14.241730 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:14.241872 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:14.241842 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:15.777309 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:15.777260 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:15.777846 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:15.777441 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:15.777846 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:15.777524 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:26:23.777503148 +0000 UTC m=+17.234987651 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:15.878978 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:15.878271 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:15.878978 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:15.878477 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 14:26:15.878978 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:15.878499 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 14:26:15.878978 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:15.878514 2580 projected.go:194] Error preparing data for projected volume kube-api-access-nfdjb for pod openshift-network-diagnostics/network-check-target-j6wgt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:15.878978 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:15.878596 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb podName:5578d992-9c91-4c3b-8499-59e3e5261890 nodeName:}" failed. No retries permitted until 2026-04-21 14:26:23.878557594 +0000 UTC m=+17.336042104 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-nfdjb" (UniqueName: "kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb") pod "network-check-target-j6wgt" (UID: "5578d992-9c91-4c3b-8499-59e3e5261890") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:16.241842 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:16.241747 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:16.242000 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:16.241887 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:16.242293 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:16.242276 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:16.242408 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:16.242387 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:18.241550 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:18.241306 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:18.242029 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:18.241306 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:18.242029 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:18.241680 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:18.242029 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:18.241783 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:20.241100 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:20.241056 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:20.241540 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:20.241056 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:20.241540 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:20.241209 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:20.241540 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:20.241237 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:22.241453 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:22.241415 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:22.241453 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:22.241460 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:22.242030 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:22.241577 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:22.242030 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:22.241674 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:23.837053 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:23.837013 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:23.837621 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:23.837134 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:23.837621 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:23.837215 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:26:39.837192952 +0000 UTC m=+33.294677452 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:23.937635 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:23.937604 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:23.937830 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:23.937758 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 14:26:23.937830 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:23.937774 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 14:26:23.937830 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:23.937784 2580 projected.go:194] Error preparing data for projected volume kube-api-access-nfdjb for pod openshift-network-diagnostics/network-check-target-j6wgt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:23.938020 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:23.937841 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb podName:5578d992-9c91-4c3b-8499-59e3e5261890 nodeName:}" failed. No retries permitted until 2026-04-21 14:26:39.937821413 +0000 UTC m=+33.395305900 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-nfdjb" (UniqueName: "kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb") pod "network-check-target-j6wgt" (UID: "5578d992-9c91-4c3b-8499-59e3e5261890") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:24.240949 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:24.240870 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:24.240949 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:24.240883 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:24.241169 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:24.240988 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:24.241169 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:24.241123 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:26.240728 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:26.240692 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:26.241150 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:26.240694 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:26.241150 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:26.240822 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:26.241150 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:26.240910 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:27.431463 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.431101 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rftw" event={"ID":"ef45360a-730f-4594-a33e-4ae5f5cd4757","Type":"ContainerStarted","Data":"932e9397bf28c53525163abcaa81bd83fe315f83c2fba9e3ad95f62fbe9295aa"} Apr 21 14:26:27.437889 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.437856 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" event={"ID":"77abbe16-9a16-46f3-9e44-0d4c056849cd","Type":"ContainerStarted","Data":"183e0d8656175f54707d41aea037b0ebc4e5178b1eabda2e8777184fd50d5de9"} Apr 21 14:26:27.440633 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.440613 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:26:27.440963 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.440939 2580 generic.go:358] "Generic (PLEG): container finished" podID="05386859-dd8b-49f4-8271-44949e102c17" containerID="6b3f3303ca1b41884a273ba49ce81fd1c4d713f775369febc628aaee4db32c7e" exitCode=1 Apr 21 14:26:27.441025 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.441001 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"dde22c649c8538ddd5862c77d174c8e4b52293e72be7ce548a24def4ae7cce45"} Apr 21 14:26:27.441080 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.441031 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"09dde4b4c057227a4efb594872cd3021b9ce6ce3f870973b857134a0af19b603"} Apr 21 14:26:27.441080 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.441041 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"3b0714721404693a0582c3f637086bd9b7b4d7bea0326383660acb19a9fdfbca"} Apr 21 14:26:27.441080 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.441049 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"7c92b08a4908ea39d64da3f6f5babb876df332b53dc848957ed6bc8c030b3f4b"} Apr 21 14:26:27.441080 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.441057 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerDied","Data":"6b3f3303ca1b41884a273ba49ce81fd1c4d713f775369febc628aaee4db32c7e"} Apr 21 14:26:27.441080 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.441067 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"577dd3fe6ab31b8a683227cd7635db55ccab688cc7bdf402938466e7730c5289"} Apr 21 14:26:27.442366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.442347 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" event={"ID":"30f945062f07b4b1388971557d0103c9","Type":"ContainerStarted","Data":"666f2f31ca503f141b337ec94d5075791f16d8e447fb2bbee6126611c77b101e"} Apr 21 14:26:27.472689 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.472643 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8rftw" podStartSLOduration=2.294674079 podStartE2EDuration="20.472629805s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.495271649 +0000 UTC m=+1.952756137" lastFinishedPulling="2026-04-21 14:26:26.673227363 +0000 UTC m=+20.130711863" observedRunningTime="2026-04-21 14:26:27.454668497 +0000 UTC m=+20.912153004" watchObservedRunningTime="2026-04-21 14:26:27.472629805 +0000 UTC m=+20.930114311" Apr 21 14:26:27.473071 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.473043 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-61.ec2.internal" podStartSLOduration=20.473034277 podStartE2EDuration="20.473034277s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:26:27.473000593 +0000 UTC m=+20.930485099" watchObservedRunningTime="2026-04-21 14:26:27.473034277 +0000 UTC m=+20.930518783" Apr 21 14:26:27.495212 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:27.495173 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-ld9pw" podStartSLOduration=2.429168779 podStartE2EDuration="20.49516085s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.480728283 +0000 UTC m=+1.938212768" lastFinishedPulling="2026-04-21 14:26:26.546720339 +0000 UTC m=+20.004204839" observedRunningTime="2026-04-21 14:26:27.494993483 +0000 UTC m=+20.952477989" watchObservedRunningTime="2026-04-21 14:26:27.49516085 +0000 UTC m=+20.952645358" Apr 21 14:26:28.241243 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.241068 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:28.241440 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.241084 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:28.241440 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:28.241337 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:28.241533 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:28.241454 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:28.446164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.446127 2580 generic.go:358] "Generic (PLEG): container finished" podID="f283d499d69cf4025cd4d0f1d7e7bb17" containerID="631a0b4d7efdf5c32b77eb235a8b66bd0b7423aed48fe5fe322cc964d4d57229" exitCode=0 Apr 21 14:26:28.446882 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.446404 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" event={"ID":"f283d499d69cf4025cd4d0f1d7e7bb17","Type":"ContainerDied","Data":"631a0b4d7efdf5c32b77eb235a8b66bd0b7423aed48fe5fe322cc964d4d57229"} Apr 21 14:26:28.448139 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.448111 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-jzct4" event={"ID":"17883801-bbff-4ed7-8ba5-44081b8c5476","Type":"ContainerStarted","Data":"666cb1946f9e2c5d23bb11d09075d5ad8749e28f7dc0ed80b117778b7edfa3e0"} Apr 21 14:26:28.449544 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.449484 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-89ch8" event={"ID":"9ed791bd-825b-4608-ad66-0304c5a34a75","Type":"ContainerStarted","Data":"fe2fbcf361e590c5910f347f9cc42c9ecbfa8d583858a0174f38ef727ab937b7"} Apr 21 14:26:28.451030 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.451007 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lmrrn" event={"ID":"ccae8641-954f-48b3-85eb-812164adcec6","Type":"ContainerStarted","Data":"8fa851c62712f6d9cbbe57f6eda4cbb0b72d01553c53e71a9907f277a6c5ea6e"} Apr 21 14:26:28.454229 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.454205 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" event={"ID":"a4e7a81b-2051-4a16-947b-2430ad7a69ec","Type":"ContainerStarted","Data":"414cd8d418715d9e47ca0087c3c2fb1baa79712b02148e6e13b85c45ae9bb7f1"} Apr 21 14:26:28.456003 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.455975 2580 generic.go:358] "Generic (PLEG): container finished" podID="29a15288-a939-4c42-b167-118380ab5462" containerID="ee0aacd72451f9f49ccd96dd5170a42c06fbdfe3ba32676fe60a1cc8a876ef7e" exitCode=0 Apr 21 14:26:28.456098 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.456052 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerDied","Data":"ee0aacd72451f9f49ccd96dd5170a42c06fbdfe3ba32676fe60a1cc8a876ef7e"} Apr 21 14:26:28.457477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.457455 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-rjjmt" event={"ID":"45cf2d21-c2f8-43b1-95f7-d3afd9014c36","Type":"ContainerStarted","Data":"a8c9e753485fe11fc06c49864d6d8bef58c06217864b57434260bb365004433f"} Apr 21 14:26:28.474273 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.474224 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lmrrn" podStartSLOduration=3.362277468 podStartE2EDuration="21.474211235s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.423375334 +0000 UTC m=+1.880859819" lastFinishedPulling="2026-04-21 14:26:26.535309085 +0000 UTC m=+19.992793586" observedRunningTime="2026-04-21 14:26:28.473625012 +0000 UTC m=+21.931109520" watchObservedRunningTime="2026-04-21 14:26:28.474211235 +0000 UTC m=+21.931695742" Apr 21 14:26:28.486969 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.486931 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-89ch8" podStartSLOduration=7.881059218 podStartE2EDuration="21.486919285s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.436155141 +0000 UTC m=+1.893639626" lastFinishedPulling="2026-04-21 14:26:22.042015206 +0000 UTC m=+15.499499693" observedRunningTime="2026-04-21 14:26:28.48633927 +0000 UTC m=+21.943823778" watchObservedRunningTime="2026-04-21 14:26:28.486919285 +0000 UTC m=+21.944403791" Apr 21 14:26:28.524494 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.524471 2580 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 21 14:26:28.530015 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.529777 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-rjjmt" podStartSLOduration=3.457723947 podStartE2EDuration="21.529760193s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.463270144 +0000 UTC m=+1.920754628" lastFinishedPulling="2026-04-21 14:26:26.535306385 +0000 UTC m=+19.992790874" observedRunningTime="2026-04-21 14:26:28.501745035 +0000 UTC m=+21.959229544" watchObservedRunningTime="2026-04-21 14:26:28.529760193 +0000 UTC m=+21.987244705" Apr 21 14:26:28.545990 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:28.545953 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-jzct4" podStartSLOduration=3.6474258649999998 podStartE2EDuration="21.54594198s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.635759746 +0000 UTC m=+2.093244237" lastFinishedPulling="2026-04-21 14:26:26.534275858 +0000 UTC m=+19.991760352" observedRunningTime="2026-04-21 14:26:28.545774279 +0000 UTC m=+22.003258786" watchObservedRunningTime="2026-04-21 14:26:28.54594198 +0000 UTC m=+22.003426486" Apr 21 14:26:29.176710 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.176626 2580 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-21T14:26:28.524491085Z","UUID":"1b0b644b-4dad-444b-b6bc-2aded9211eff","Handler":null,"Name":"","Endpoint":""} Apr 21 14:26:29.179799 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.179776 2580 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 21 14:26:29.179899 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.179809 2580 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 21 14:26:29.463937 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.463724 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" event={"ID":"a4e7a81b-2051-4a16-947b-2430ad7a69ec","Type":"ContainerStarted","Data":"ed8008bfe32b98698ac7798fd911294f5ab8d9bf9c1b383e8aa48f3e67da0f5d"} Apr 21 14:26:29.467759 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.467733 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:26:29.468305 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.468270 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"9e7c6046b44b8652a8975fd452d762fd7567a43e666f36d9b70893b038070ece"} Apr 21 14:26:29.470484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.470458 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" event={"ID":"f283d499d69cf4025cd4d0f1d7e7bb17","Type":"ContainerStarted","Data":"00e52b3de71e766cff23d7f556b3309a1da983a11a424d311a9c3924b4d3b239"} Apr 21 14:26:29.486882 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:29.486833 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-61.ec2.internal" podStartSLOduration=22.486817185 podStartE2EDuration="22.486817185s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:26:29.486667728 +0000 UTC m=+22.944152259" watchObservedRunningTime="2026-04-21 14:26:29.486817185 +0000 UTC m=+22.944301691" Apr 21 14:26:30.241689 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:30.241656 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:30.241918 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:30.241656 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:30.241918 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:30.241755 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:30.241918 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:30.241854 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:30.475004 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:30.474630 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" event={"ID":"a4e7a81b-2051-4a16-947b-2430ad7a69ec","Type":"ContainerStarted","Data":"c4d90c4579526b820e1f2da0f963a5ee1a8473473bedbe6f3d406d067e1e23e8"} Apr 21 14:26:31.479173 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:31.479141 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:26:31.479556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:31.479528 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"3dda621176eea1ac8c9ed4c25f77fa09565de03b7b625753f7fd957b038203b5"} Apr 21 14:26:31.480057 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:31.480043 2580 scope.go:117] "RemoveContainer" containerID="6b3f3303ca1b41884a273ba49ce81fd1c4d713f775369febc628aaee4db32c7e" Apr 21 14:26:31.513744 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:31.513703 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5g29d" podStartSLOduration=3.0712212819999998 podStartE2EDuration="24.513687317s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.286307173 +0000 UTC m=+1.743791662" lastFinishedPulling="2026-04-21 14:26:29.728773199 +0000 UTC m=+23.186257697" observedRunningTime="2026-04-21 14:26:30.490917179 +0000 UTC m=+23.948401687" watchObservedRunningTime="2026-04-21 14:26:31.513687317 +0000 UTC m=+24.971171847" Apr 21 14:26:32.241582 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.241536 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:32.241729 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.241547 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:32.241729 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:32.241659 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:32.241810 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:32.241726 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:32.485559 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.485523 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:26:32.486059 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.485914 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" event={"ID":"05386859-dd8b-49f4-8271-44949e102c17","Type":"ContainerStarted","Data":"d86b337798dc0239abaeddefae286d2e3c08979212cebfc9e1b3c12af1ed6339"} Apr 21 14:26:32.486238 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.486218 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:32.486337 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.486326 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:32.486429 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.486416 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:32.504905 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.504833 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:32.509101 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.509037 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:26:32.516268 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:32.516228 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" podStartSLOduration=7.414602201 podStartE2EDuration="25.516216897s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.614351711 +0000 UTC m=+2.071836196" lastFinishedPulling="2026-04-21 14:26:26.715966388 +0000 UTC m=+20.173450892" observedRunningTime="2026-04-21 14:26:32.515960528 +0000 UTC m=+25.973445037" watchObservedRunningTime="2026-04-21 14:26:32.516216897 +0000 UTC m=+25.973701404" Apr 21 14:26:33.333598 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:33.333538 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:33.334231 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:33.334207 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:33.416471 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:33.416440 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-j6wgt"] Apr 21 14:26:33.416791 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:33.416769 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:33.416926 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:33.416883 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:33.422446 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:33.422414 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fl65d"] Apr 21 14:26:33.422556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:33.422528 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:33.422654 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:33.422635 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:34.585214 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:34.585181 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:34.585973 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:34.585304 2580 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 21 14:26:34.585973 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:34.585840 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-jzct4" Apr 21 14:26:35.241227 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:35.241187 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:35.241465 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:35.241187 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:35.241465 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:35.241293 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:35.241465 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:35.241364 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:35.493047 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:35.492962 2580 generic.go:358] "Generic (PLEG): container finished" podID="29a15288-a939-4c42-b167-118380ab5462" containerID="40d8cce51794bc7c6e127394d62988dc187acf5e4e0913d54af349fc46830fa1" exitCode=0 Apr 21 14:26:35.493182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:35.493043 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerDied","Data":"40d8cce51794bc7c6e127394d62988dc187acf5e4e0913d54af349fc46830fa1"} Apr 21 14:26:36.496689 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:36.496496 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerStarted","Data":"c611b64746427dc62edfe71c99925fb93fb310fb780a47f9acfe8948f649f3a4"} Apr 21 14:26:37.242925 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:37.242893 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:37.243127 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:37.243018 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:37.243127 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:37.243079 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:37.243233 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:37.243177 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:37.499662 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:37.499559 2580 generic.go:358] "Generic (PLEG): container finished" podID="29a15288-a939-4c42-b167-118380ab5462" containerID="c611b64746427dc62edfe71c99925fb93fb310fb780a47f9acfe8948f649f3a4" exitCode=0 Apr 21 14:26:37.499662 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:37.499594 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerDied","Data":"c611b64746427dc62edfe71c99925fb93fb310fb780a47f9acfe8948f649f3a4"} Apr 21 14:26:39.240878 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.240840 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:39.241355 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.240851 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:39.241355 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.240954 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fl65d" podUID="2f89a533-2e7f-4655-bd1f-cf6e02cc46ed" Apr 21 14:26:39.241355 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.241016 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-j6wgt" podUID="5578d992-9c91-4c3b-8499-59e3e5261890" Apr 21 14:26:39.505765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.505694 2580 generic.go:358] "Generic (PLEG): container finished" podID="29a15288-a939-4c42-b167-118380ab5462" containerID="37438de8c8f02cd63175d009ebe81d094f565a0783b640ae95173c8d06d0b1d7" exitCode=0 Apr 21 14:26:39.505765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.505751 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerDied","Data":"37438de8c8f02cd63175d009ebe81d094f565a0783b640ae95173c8d06d0b1d7"} Apr 21 14:26:39.850911 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.850816 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:39.851081 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.850961 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:39.851081 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.851034 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:27:11.851015233 +0000 UTC m=+65.308499732 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 14:26:39.881625 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.881596 2580 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-61.ec2.internal" event="NodeReady" Apr 21 14:26:39.881765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.881710 2580 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 21 14:26:39.929842 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.929815 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dx8mq"] Apr 21 14:26:39.951779 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.951742 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:39.951922 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.951880 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 14:26:39.951922 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.951906 2580 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 14:26:39.951922 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.951919 2580 projected.go:194] Error preparing data for projected volume kube-api-access-nfdjb for pod openshift-network-diagnostics/network-check-target-j6wgt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:39.952141 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:39.951980 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb podName:5578d992-9c91-4c3b-8499-59e3e5261890 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:11.951963177 +0000 UTC m=+65.409447679 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-nfdjb" (UniqueName: "kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb") pod "network-check-target-j6wgt" (UID: "5578d992-9c91-4c3b-8499-59e3e5261890") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 14:26:39.964217 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.964192 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-tfksq"] Apr 21 14:26:39.964352 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.964336 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:39.966534 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.966515 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 21 14:26:39.966671 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.966515 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 21 14:26:39.966671 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.966547 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-88srx\"" Apr 21 14:26:39.987395 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.987372 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dx8mq"] Apr 21 14:26:39.987395 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.987396 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tfksq"] Apr 21 14:26:39.987538 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.987484 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:39.989661 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.989643 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 21 14:26:39.989781 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.989771 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-pqf56\"" Apr 21 14:26:39.989847 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.989803 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 21 14:26:39.989847 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:39.989811 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 21 14:26:40.053100 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.053064 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:40.053285 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.053117 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9dw9\" (UniqueName: \"kubernetes.io/projected/5badb5d3-1012-4fbf-81cc-b5190659075e-kube-api-access-j9dw9\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:40.053285 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.053217 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.053285 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.053269 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/afea4c93-a952-4362-8775-179f3919eecc-tmp-dir\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.053462 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.053321 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afea4c93-a952-4362-8775-179f3919eecc-config-volume\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.053462 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.053361 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7nlf\" (UniqueName: \"kubernetes.io/projected/afea4c93-a952-4362-8775-179f3919eecc-kube-api-access-s7nlf\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.154007 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.153929 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:40.154007 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.153973 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-j9dw9\" (UniqueName: \"kubernetes.io/projected/5badb5d3-1012-4fbf-81cc-b5190659075e-kube-api-access-j9dw9\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.154017 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.154054 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/afea4c93-a952-4362-8775-179f3919eecc-tmp-dir\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.154095 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afea4c93-a952-4362-8775-179f3919eecc-config-volume\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.154106 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.154119 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-s7nlf\" (UniqueName: \"kubernetes.io/projected/afea4c93-a952-4362-8775-179f3919eecc-kube-api-access-s7nlf\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.154172 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.154180 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:26:40.654159193 +0000 UTC m=+34.111643723 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:26:40.154245 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.154242 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:26:40.654222822 +0000 UTC m=+34.111707320 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:26:40.154581 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.154463 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/afea4c93-a952-4362-8775-179f3919eecc-tmp-dir\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.154662 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.154642 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afea4c93-a952-4362-8775-179f3919eecc-config-volume\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.164742 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.164641 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7nlf\" (UniqueName: \"kubernetes.io/projected/afea4c93-a952-4362-8775-179f3919eecc-kube-api-access-s7nlf\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.164851 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.164787 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9dw9\" (UniqueName: \"kubernetes.io/projected/5badb5d3-1012-4fbf-81cc-b5190659075e-kube-api-access-j9dw9\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:40.657636 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.657597 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:40.658072 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:40.657657 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:40.658072 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.657760 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:26:40.658072 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.657781 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:26:40.658072 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.657833 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:26:41.657810572 +0000 UTC m=+35.115295060 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:26:40.658072 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:40.657849 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:26:41.65784258 +0000 UTC m=+35.115327065 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:26:41.241216 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.241184 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:26:41.241432 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.241185 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:26:41.245159 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.245006 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 14:26:41.245159 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.245019 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 21 14:26:41.245159 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.245028 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 21 14:26:41.245159 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.245066 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-fd928\"" Apr 21 14:26:41.245159 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.245022 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-9wlmf\"" Apr 21 14:26:41.665867 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.665829 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:41.666295 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:41.665902 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:41.666295 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:41.665992 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:26:41.666295 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:41.666019 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:26:41.666295 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:41.666093 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:26:43.666074421 +0000 UTC m=+37.123558909 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:26:41.666295 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:41.666118 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:26:43.666108255 +0000 UTC m=+37.123592743 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:26:43.682652 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:43.682610 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:43.683097 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:43.682703 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:43.683097 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:43.682759 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:26:43.683097 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:43.682804 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:26:43.683097 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:43.682837 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:26:47.682815304 +0000 UTC m=+41.140299801 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:26:43.683097 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:43.682855 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:26:47.682847132 +0000 UTC m=+41.140331619 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:26:46.520919 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:46.520876 2580 generic.go:358] "Generic (PLEG): container finished" podID="29a15288-a939-4c42-b167-118380ab5462" containerID="4dea505370d5383e47994cd03325ed69408b23e28891d68a3791544874794495" exitCode=0 Apr 21 14:26:46.521276 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:46.520946 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerDied","Data":"4dea505370d5383e47994cd03325ed69408b23e28891d68a3791544874794495"} Apr 21 14:26:47.525290 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:47.525253 2580 generic.go:358] "Generic (PLEG): container finished" podID="29a15288-a939-4c42-b167-118380ab5462" containerID="493b08d997517444ab88cbb9cfc79a796878f5b007e05720f3e421119960dab4" exitCode=0 Apr 21 14:26:47.525704 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:47.525321 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerDied","Data":"493b08d997517444ab88cbb9cfc79a796878f5b007e05720f3e421119960dab4"} Apr 21 14:26:47.716135 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:47.716091 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:47.716266 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:47.716153 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:47.716266 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:47.716239 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:26:47.716349 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:47.716299 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:26:55.716283748 +0000 UTC m=+49.173768233 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:26:47.716349 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:47.716238 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:26:47.716416 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:47.716389 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:26:55.716371368 +0000 UTC m=+49.173855867 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:26:48.529874 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:48.529833 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" event={"ID":"29a15288-a939-4c42-b167-118380ab5462","Type":"ContainerStarted","Data":"6fc49427982c8cdafbe9569aae8305bacc25a2942347ab1c390af6df72fcad33"} Apr 21 14:26:48.555035 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:48.554974 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-sz9sg" podStartSLOduration=4.681443026 podStartE2EDuration="41.554955209s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:26:08.517450218 +0000 UTC m=+1.974934703" lastFinishedPulling="2026-04-21 14:26:45.390962397 +0000 UTC m=+38.848446886" observedRunningTime="2026-04-21 14:26:48.552776448 +0000 UTC m=+42.010261006" watchObservedRunningTime="2026-04-21 14:26:48.554955209 +0000 UTC m=+42.012439717" Apr 21 14:26:55.770263 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:55.770213 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:26:55.770839 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:26:55.770279 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:26:55.770839 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:55.770372 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:26:55.770839 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:55.770417 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:26:55.770839 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:55.770445 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:27:11.770428688 +0000 UTC m=+65.227913172 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:26:55.770839 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:26:55.770462 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:27:11.770451593 +0000 UTC m=+65.227936078 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:27:04.514551 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:04.514523 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ts88z" Apr 21 14:27:11.781201 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:11.781156 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:27:11.781686 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:11.781238 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:27:11.781686 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:11.781330 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:27:11.781686 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:11.781356 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:27:11.781686 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:11.781409 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:27:43.781390965 +0000 UTC m=+97.238875454 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:27:11.781686 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:11.781426 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:27:43.781418874 +0000 UTC m=+97.238903359 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:27:11.882412 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:11.882370 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:27:11.884865 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:11.884845 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 14:27:11.893036 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:11.893015 2580 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 21 14:27:11.893136 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:11.893076 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs podName:2f89a533-2e7f-4655-bd1f-cf6e02cc46ed nodeName:}" failed. No retries permitted until 2026-04-21 14:28:15.89306005 +0000 UTC m=+129.350544534 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs") pod "network-metrics-daemon-fl65d" (UID: "2f89a533-2e7f-4655-bd1f-cf6e02cc46ed") : secret "metrics-daemon-secret" not found Apr 21 14:27:11.983453 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:11.983409 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:27:11.985994 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:11.985974 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 21 14:27:11.995844 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:11.995818 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 21 14:27:12.007467 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:12.007441 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfdjb\" (UniqueName: \"kubernetes.io/projected/5578d992-9c91-4c3b-8499-59e3e5261890-kube-api-access-nfdjb\") pod \"network-check-target-j6wgt\" (UID: \"5578d992-9c91-4c3b-8499-59e3e5261890\") " pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:27:12.162948 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:12.162862 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-fd928\"" Apr 21 14:27:12.170875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:12.170849 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:27:12.302096 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:12.302064 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-j6wgt"] Apr 21 14:27:12.575130 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:12.575080 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-j6wgt" event={"ID":"5578d992-9c91-4c3b-8499-59e3e5261890","Type":"ContainerStarted","Data":"f222d79affdb2712a64155c5c4603e9926dec760eb031de52070462fb0c06aa1"} Apr 21 14:27:15.582916 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:15.582878 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-j6wgt" event={"ID":"5578d992-9c91-4c3b-8499-59e3e5261890","Type":"ContainerStarted","Data":"66c3cafb04a895e3bf7df06906776ddb30bbbaca8e475144b29048ee69f2d063"} Apr 21 14:27:15.583319 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:15.583042 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:27:15.600260 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:15.600205 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-j6wgt" podStartSLOduration=65.901201179 podStartE2EDuration="1m8.600192315s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:27:12.307871028 +0000 UTC m=+65.765355528" lastFinishedPulling="2026-04-21 14:27:15.006862168 +0000 UTC m=+68.464346664" observedRunningTime="2026-04-21 14:27:15.599775258 +0000 UTC m=+69.057259764" watchObservedRunningTime="2026-04-21 14:27:15.600192315 +0000 UTC m=+69.057676821" Apr 21 14:27:35.623771 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.623732 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df"] Apr 21 14:27:35.627956 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.627932 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-operator-585dfdc468-ckmd7"] Apr 21 14:27:35.628097 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.628079 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.630541 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.630525 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.632277 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632245 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-tls\"" Apr 21 14:27:35.632277 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632262 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 21 14:27:35.632277 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632274 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 21 14:27:35.632464 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632291 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-r7pvz\"" Apr 21 14:27:35.632464 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632398 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemetry-config\"" Apr 21 14:27:35.632803 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632775 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 21 14:27:35.632938 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632804 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"service-ca-bundle\"" Apr 21 14:27:35.632938 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632811 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 21 14:27:35.632938 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.632896 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"openshift-insights-serving-cert\"" Apr 21 14:27:35.633124 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.633095 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"operator-dockercfg-ds462\"" Apr 21 14:27:35.634917 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.634899 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df"] Apr 21 14:27:35.637939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.637918 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"trusted-ca-bundle\"" Apr 21 14:27:35.642386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642359 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c20781a-bb09-4f47-8dcd-d4d713caac69-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.642476 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642410 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c20781a-bb09-4f47-8dcd-d4d713caac69-service-ca-bundle\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.642476 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642426 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-ckmd7"] Apr 21 14:27:35.642476 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642443 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c20781a-bb09-4f47-8dcd-d4d713caac69-serving-cert\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.642682 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642473 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89sqn\" (UniqueName: \"kubernetes.io/projected/f355a71a-6279-4fa0-9123-6a056ee88f72-kube-api-access-89sqn\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.642682 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642608 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4c20781a-bb09-4f47-8dcd-d4d713caac69-tmp\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.642682 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642638 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.642682 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642663 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/4c20781a-bb09-4f47-8dcd-d4d713caac69-snapshots\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.642874 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642683 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/f355a71a-6279-4fa0-9123-6a056ee88f72-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.642874 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.642727 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f56t\" (UniqueName: \"kubernetes.io/projected/4c20781a-bb09-4f47-8dcd-d4d713caac69-kube-api-access-7f56t\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.726721 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.726685 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-6d6c54d4dd-mb7h8"] Apr 21 14:27:35.729644 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.729629 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.731949 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.731920 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Apr 21 14:27:35.732077 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.731925 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Apr 21 14:27:35.732077 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.732041 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-x64rj\"" Apr 21 14:27:35.732488 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.732472 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Apr 21 14:27:35.732540 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.732510 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Apr 21 14:27:35.732939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.732924 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Apr 21 14:27:35.733019 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.733003 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"default-ingress-cert\"" Apr 21 14:27:35.742871 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.742847 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/router-default-6d6c54d4dd-mb7h8"] Apr 21 14:27:35.743035 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743017 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7f56t\" (UniqueName: \"kubernetes.io/projected/4c20781a-bb09-4f47-8dcd-d4d713caac69-kube-api-access-7f56t\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.743089 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743046 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-default-certificate\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.743089 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743066 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-stats-auth\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.743236 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743224 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c20781a-bb09-4f47-8dcd-d4d713caac69-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.743275 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743243 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.743275 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743258 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.743357 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743277 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c20781a-bb09-4f47-8dcd-d4d713caac69-service-ca-bundle\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.743357 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743293 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c20781a-bb09-4f47-8dcd-d4d713caac69-serving-cert\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.743357 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743332 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-89sqn\" (UniqueName: \"kubernetes.io/projected/f355a71a-6279-4fa0-9123-6a056ee88f72-kube-api-access-89sqn\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.743486 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743456 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4c20781a-bb09-4f47-8dcd-d4d713caac69-tmp\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.743486 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743479 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.743601 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743501 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/4c20781a-bb09-4f47-8dcd-d4d713caac69-snapshots\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.743601 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743530 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/f355a71a-6279-4fa0-9123-6a056ee88f72-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.743601 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.743583 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5gr4\" (UniqueName: \"kubernetes.io/projected/0284ee69-97a8-4c33-98ee-1bb75e7f461b-kube-api-access-v5gr4\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.743872 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:35.743852 2580 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:35.744033 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.744010 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c20781a-bb09-4f47-8dcd-d4d713caac69-service-ca-bundle\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.744138 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:35.744021 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls podName:f355a71a-6279-4fa0-9123-6a056ee88f72 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:36.244001789 +0000 UTC m=+89.701486288 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-pn2df" (UID: "f355a71a-6279-4fa0-9123-6a056ee88f72") : secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:35.744366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.744322 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/f355a71a-6279-4fa0-9123-6a056ee88f72-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.744449 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.744434 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4c20781a-bb09-4f47-8dcd-d4d713caac69-tmp\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.744713 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.744691 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/4c20781a-bb09-4f47-8dcd-d4d713caac69-snapshots\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.744914 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.744894 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c20781a-bb09-4f47-8dcd-d4d713caac69-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.745869 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.745853 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c20781a-bb09-4f47-8dcd-d4d713caac69-serving-cert\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.757115 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.757094 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f56t\" (UniqueName: \"kubernetes.io/projected/4c20781a-bb09-4f47-8dcd-d4d713caac69-kube-api-access-7f56t\") pod \"insights-operator-585dfdc468-ckmd7\" (UID: \"4c20781a-bb09-4f47-8dcd-d4d713caac69\") " pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:35.759855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.759823 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-89sqn\" (UniqueName: \"kubernetes.io/projected/f355a71a-6279-4fa0-9123-6a056ee88f72-kube-api-access-89sqn\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:35.844724 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.844681 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v5gr4\" (UniqueName: \"kubernetes.io/projected/0284ee69-97a8-4c33-98ee-1bb75e7f461b-kube-api-access-v5gr4\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.844924 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.844743 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-default-certificate\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.844924 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.844767 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-stats-auth\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.845043 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.844915 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.845043 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.844965 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.845158 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:35.845061 2580 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 14:27:35.845158 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:35.845076 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:36.345059273 +0000 UTC m=+89.802543791 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : configmap references non-existent config key: service-ca.crt Apr 21 14:27:35.845261 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:35.845161 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:36.345143161 +0000 UTC m=+89.802627650 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : secret "router-metrics-certs-default" not found Apr 21 14:27:35.847269 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.847246 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-stats-auth\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.847372 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.847328 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-default-certificate\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.857841 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.857819 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5gr4\" (UniqueName: \"kubernetes.io/projected/0284ee69-97a8-4c33-98ee-1bb75e7f461b-kube-api-access-v5gr4\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:35.945123 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:35.945088 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-ckmd7" Apr 21 14:27:36.062168 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:36.062139 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-ckmd7"] Apr 21 14:27:36.064972 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:27:36.064948 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c20781a_bb09_4f47_8dcd_d4d713caac69.slice/crio-3b58a99dcdd67f7ec1b9f8987942f520b84faf20149dfeaa1998bfac4f7485d3 WatchSource:0}: Error finding container 3b58a99dcdd67f7ec1b9f8987942f520b84faf20149dfeaa1998bfac4f7485d3: Status 404 returned error can't find the container with id 3b58a99dcdd67f7ec1b9f8987942f520b84faf20149dfeaa1998bfac4f7485d3 Apr 21 14:27:36.248472 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:36.248378 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:36.248640 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:36.248527 2580 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:36.248640 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:36.248629 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls podName:f355a71a-6279-4fa0-9123-6a056ee88f72 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:37.248608979 +0000 UTC m=+90.706093469 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-pn2df" (UID: "f355a71a-6279-4fa0-9123-6a056ee88f72") : secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:36.349666 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:36.349627 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:36.349666 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:36.349669 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:36.349905 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:36.349764 2580 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 14:27:36.349905 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:36.349788 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:37.349768755 +0000 UTC m=+90.807253249 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : configmap references non-existent config key: service-ca.crt Apr 21 14:27:36.349905 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:36.349823 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:37.349807731 +0000 UTC m=+90.807292231 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : secret "router-metrics-certs-default" not found Apr 21 14:27:36.626544 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:36.626460 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-ckmd7" event={"ID":"4c20781a-bb09-4f47-8dcd-d4d713caac69","Type":"ContainerStarted","Data":"3b58a99dcdd67f7ec1b9f8987942f520b84faf20149dfeaa1998bfac4f7485d3"} Apr 21 14:27:37.256196 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.256161 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:37.256396 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:37.256314 2580 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:37.256396 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:37.256387 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls podName:f355a71a-6279-4fa0-9123-6a056ee88f72 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:39.256371478 +0000 UTC m=+92.713855965 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-pn2df" (UID: "f355a71a-6279-4fa0-9123-6a056ee88f72") : secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:37.356988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.356945 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:37.356988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.356985 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:37.357199 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:37.357088 2580 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 14:27:37.357199 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:37.357106 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:39.357088609 +0000 UTC m=+92.814573094 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : configmap references non-existent config key: service-ca.crt Apr 21 14:27:37.357199 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:37.357133 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:39.357120144 +0000 UTC m=+92.814604628 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : secret "router-metrics-certs-default" not found Apr 21 14:27:37.541619 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.541519 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf"] Apr 21 14:27:37.545787 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.545764 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" Apr 21 14:27:37.548125 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.548096 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-storage-operator\"/\"kube-root-ca.crt\"" Apr 21 14:27:37.548492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.548464 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-storage-operator\"/\"openshift-service-ca.crt\"" Apr 21 14:27:37.548619 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.548469 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-storage-operator\"/\"volume-data-source-validator-dockercfg-b7n2w\"" Apr 21 14:27:37.551521 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.551496 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf"] Apr 21 14:27:37.560588 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.560535 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6hw\" (UniqueName: \"kubernetes.io/projected/898bbae2-8a55-44cc-b13f-5ab0167554e4-kube-api-access-zp6hw\") pod \"volume-data-source-validator-7c6cbb6c87-2mnrf\" (UID: \"898bbae2-8a55-44cc-b13f-5ab0167554e4\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" Apr 21 14:27:37.640161 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.640121 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh"] Apr 21 14:27:37.643211 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.643191 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:37.647370 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.647182 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-zx6fl\"" Apr 21 14:27:37.647370 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.647182 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Apr 21 14:27:37.647536 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.647373 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Apr 21 14:27:37.647536 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.647510 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Apr 21 14:27:37.656906 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.656881 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh"] Apr 21 14:27:37.661173 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.661149 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6hw\" (UniqueName: \"kubernetes.io/projected/898bbae2-8a55-44cc-b13f-5ab0167554e4-kube-api-access-zp6hw\") pod \"volume-data-source-validator-7c6cbb6c87-2mnrf\" (UID: \"898bbae2-8a55-44cc-b13f-5ab0167554e4\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" Apr 21 14:27:37.661312 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.661235 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:37.661312 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.661296 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzwhn\" (UniqueName: \"kubernetes.io/projected/4402dc96-86f0-45e6-b15c-7aace6e551be-kube-api-access-xzwhn\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:37.671305 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.671268 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6hw\" (UniqueName: \"kubernetes.io/projected/898bbae2-8a55-44cc-b13f-5ab0167554e4-kube-api-access-zp6hw\") pod \"volume-data-source-validator-7c6cbb6c87-2mnrf\" (UID: \"898bbae2-8a55-44cc-b13f-5ab0167554e4\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" Apr 21 14:27:37.762156 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.762114 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xzwhn\" (UniqueName: \"kubernetes.io/projected/4402dc96-86f0-45e6-b15c-7aace6e551be-kube-api-access-xzwhn\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:37.762328 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.762248 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:37.762369 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:37.762355 2580 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 14:27:37.762432 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:37.762420 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls podName:4402dc96-86f0-45e6-b15c-7aace6e551be nodeName:}" failed. No retries permitted until 2026-04-21 14:27:38.26240011 +0000 UTC m=+91.719884600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-fjvbh" (UID: "4402dc96-86f0-45e6-b15c-7aace6e551be") : secret "samples-operator-tls" not found Apr 21 14:27:37.772054 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.772025 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzwhn\" (UniqueName: \"kubernetes.io/projected/4402dc96-86f0-45e6-b15c-7aace6e551be-kube-api-access-xzwhn\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:37.862005 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:37.861919 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" Apr 21 14:27:38.068474 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.068442 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf"] Apr 21 14:27:38.073058 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:27:38.073027 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod898bbae2_8a55_44cc_b13f_5ab0167554e4.slice/crio-1ee9d0321c0406815ab77ee5f8f8108a275d04ad7641b0363fc6effc5bd73480 WatchSource:0}: Error finding container 1ee9d0321c0406815ab77ee5f8f8108a275d04ad7641b0363fc6effc5bd73480: Status 404 returned error can't find the container with id 1ee9d0321c0406815ab77ee5f8f8108a275d04ad7641b0363fc6effc5bd73480 Apr 21 14:27:38.265931 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.265897 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:38.266124 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:38.266051 2580 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 14:27:38.266214 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:38.266144 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls podName:4402dc96-86f0-45e6-b15c-7aace6e551be nodeName:}" failed. No retries permitted until 2026-04-21 14:27:39.266123786 +0000 UTC m=+92.723608283 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-fjvbh" (UID: "4402dc96-86f0-45e6-b15c-7aace6e551be") : secret "samples-operator-tls" not found Apr 21 14:27:38.510404 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.510369 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-qjbtz"] Apr 21 14:27:38.513187 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.513171 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.515552 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.515529 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Apr 21 14:27:38.515784 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.515768 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-w62cc\"" Apr 21 14:27:38.515920 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.515898 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Apr 21 14:27:38.516177 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.516125 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Apr 21 14:27:38.516478 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.516457 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Apr 21 14:27:38.523491 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.523471 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-qjbtz"] Apr 21 14:27:38.523628 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.523608 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Apr 21 14:27:38.568344 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.568302 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c4a98bb-ce18-4746-b5b2-979619d26ebc-config\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.568538 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.568376 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxfsz\" (UniqueName: \"kubernetes.io/projected/1c4a98bb-ce18-4746-b5b2-979619d26ebc-kube-api-access-zxfsz\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.568538 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.568421 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c4a98bb-ce18-4746-b5b2-979619d26ebc-trusted-ca\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.568538 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.568465 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c4a98bb-ce18-4746-b5b2-979619d26ebc-serving-cert\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.631889 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.631855 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-ckmd7" event={"ID":"4c20781a-bb09-4f47-8dcd-d4d713caac69","Type":"ContainerStarted","Data":"abccd478d277afe4e0461e00ec21f6821ac0e3c69942deae55c86efcc9036296"} Apr 21 14:27:38.633392 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.633360 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" event={"ID":"898bbae2-8a55-44cc-b13f-5ab0167554e4","Type":"ContainerStarted","Data":"1ee9d0321c0406815ab77ee5f8f8108a275d04ad7641b0363fc6effc5bd73480"} Apr 21 14:27:38.653674 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.653630 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-operator-585dfdc468-ckmd7" podStartSLOduration=1.7274015249999999 podStartE2EDuration="3.653616117s" podCreationTimestamp="2026-04-21 14:27:35 +0000 UTC" firstStartedPulling="2026-04-21 14:27:36.066696769 +0000 UTC m=+89.524181255" lastFinishedPulling="2026-04-21 14:27:37.992911359 +0000 UTC m=+91.450395847" observedRunningTime="2026-04-21 14:27:38.652655128 +0000 UTC m=+92.110139641" watchObservedRunningTime="2026-04-21 14:27:38.653616117 +0000 UTC m=+92.111100624" Apr 21 14:27:38.669471 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.669431 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c4a98bb-ce18-4746-b5b2-979619d26ebc-config\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.669629 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.669500 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zxfsz\" (UniqueName: \"kubernetes.io/projected/1c4a98bb-ce18-4746-b5b2-979619d26ebc-kube-api-access-zxfsz\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.669629 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.669578 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c4a98bb-ce18-4746-b5b2-979619d26ebc-trusted-ca\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.669754 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.669659 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c4a98bb-ce18-4746-b5b2-979619d26ebc-serving-cert\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.670124 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.670100 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c4a98bb-ce18-4746-b5b2-979619d26ebc-config\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.671465 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.671440 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c4a98bb-ce18-4746-b5b2-979619d26ebc-trusted-ca\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.673240 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.672995 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c4a98bb-ce18-4746-b5b2-979619d26ebc-serving-cert\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.679689 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.679665 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxfsz\" (UniqueName: \"kubernetes.io/projected/1c4a98bb-ce18-4746-b5b2-979619d26ebc-kube-api-access-zxfsz\") pod \"console-operator-9d4b6777b-qjbtz\" (UID: \"1c4a98bb-ce18-4746-b5b2-979619d26ebc\") " pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.823739 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.823658 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:38.988825 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:38.988755 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-qjbtz"] Apr 21 14:27:39.240339 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:27:39.240304 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c4a98bb_ce18_4746_b5b2_979619d26ebc.slice/crio-a9e280b23996cb58cd313afd8abc429c68449f3e00478b4d3a9411686f0363c9 WatchSource:0}: Error finding container a9e280b23996cb58cd313afd8abc429c68449f3e00478b4d3a9411686f0363c9: Status 404 returned error can't find the container with id a9e280b23996cb58cd313afd8abc429c68449f3e00478b4d3a9411686f0363c9 Apr 21 14:27:39.274324 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:39.274011 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:39.274324 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:39.274064 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:39.274324 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:39.274176 2580 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:39.274324 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:39.274256 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls podName:f355a71a-6279-4fa0-9123-6a056ee88f72 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:43.274234508 +0000 UTC m=+96.731719006 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-pn2df" (UID: "f355a71a-6279-4fa0-9123-6a056ee88f72") : secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:39.274324 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:39.274266 2580 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 14:27:39.274690 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:39.274345 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls podName:4402dc96-86f0-45e6-b15c-7aace6e551be nodeName:}" failed. No retries permitted until 2026-04-21 14:27:41.274329025 +0000 UTC m=+94.731813518 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-fjvbh" (UID: "4402dc96-86f0-45e6-b15c-7aace6e551be") : secret "samples-operator-tls" not found Apr 21 14:27:39.376928 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:39.376892 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:39.377098 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:39.376943 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:39.377098 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:39.377061 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:43.377039095 +0000 UTC m=+96.834523598 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : configmap references non-existent config key: service-ca.crt Apr 21 14:27:39.377216 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:39.377112 2580 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 14:27:39.377216 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:39.377158 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:43.377145388 +0000 UTC m=+96.834629873 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : secret "router-metrics-certs-default" not found Apr 21 14:27:39.636636 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:39.636539 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" event={"ID":"898bbae2-8a55-44cc-b13f-5ab0167554e4","Type":"ContainerStarted","Data":"f0f3c48c4e81cbaf92bd2a199dbb37e74ccf189a8036a613cb16783b2eaa92f5"} Apr 21 14:27:39.637558 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:39.637531 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" event={"ID":"1c4a98bb-ce18-4746-b5b2-979619d26ebc","Type":"ContainerStarted","Data":"a9e280b23996cb58cd313afd8abc429c68449f3e00478b4d3a9411686f0363c9"} Apr 21 14:27:39.653248 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:39.653204 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-2mnrf" podStartSLOduration=1.435361028 podStartE2EDuration="2.653191744s" podCreationTimestamp="2026-04-21 14:27:37 +0000 UTC" firstStartedPulling="2026-04-21 14:27:38.074963805 +0000 UTC m=+91.532448294" lastFinishedPulling="2026-04-21 14:27:39.292794522 +0000 UTC m=+92.750279010" observedRunningTime="2026-04-21 14:27:39.652247886 +0000 UTC m=+93.109732393" watchObservedRunningTime="2026-04-21 14:27:39.653191744 +0000 UTC m=+93.110676250" Apr 21 14:27:41.297053 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:41.296966 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:41.297442 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:41.297115 2580 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 14:27:41.297442 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:41.297187 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls podName:4402dc96-86f0-45e6-b15c-7aace6e551be nodeName:}" failed. No retries permitted until 2026-04-21 14:27:45.297169901 +0000 UTC m=+98.754654386 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-fjvbh" (UID: "4402dc96-86f0-45e6-b15c-7aace6e551be") : secret "samples-operator-tls" not found Apr 21 14:27:41.475588 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:41.475541 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-lmrrn_ccae8641-954f-48b3-85eb-812164adcec6/dns-node-resolver/0.log" Apr 21 14:27:41.643782 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:41.643702 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/0.log" Apr 21 14:27:41.643782 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:41.643741 2580 generic.go:358] "Generic (PLEG): container finished" podID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" containerID="f810df9e65b4e91e7e06df2ffece6b708b9f8285659989129bfa60e47c54d6e6" exitCode=255 Apr 21 14:27:41.643958 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:41.643774 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" event={"ID":"1c4a98bb-ce18-4746-b5b2-979619d26ebc","Type":"ContainerDied","Data":"f810df9e65b4e91e7e06df2ffece6b708b9f8285659989129bfa60e47c54d6e6"} Apr 21 14:27:41.644020 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:41.644006 2580 scope.go:117] "RemoveContainer" containerID="f810df9e65b4e91e7e06df2ffece6b708b9f8285659989129bfa60e47c54d6e6" Apr 21 14:27:42.076740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.076716 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-89ch8_9ed791bd-825b-4608-ad66-0304c5a34a75/node-ca/0.log" Apr 21 14:27:42.550368 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.550332 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd"] Apr 21 14:27:42.553119 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.553103 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.556066 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.556043 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Apr 21 14:27:42.556177 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.556071 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Apr 21 14:27:42.556812 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.556793 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-9dgvg\"" Apr 21 14:27:42.556812 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.556809 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Apr 21 14:27:42.556956 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.556800 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Apr 21 14:27:42.563737 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.563714 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd"] Apr 21 14:27:42.608764 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.608730 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07d5eb46-986b-4f0e-9f2b-6db3b006c975-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.608922 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.608767 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d69kz\" (UniqueName: \"kubernetes.io/projected/07d5eb46-986b-4f0e-9f2b-6db3b006c975-kube-api-access-d69kz\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.608922 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.608891 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07d5eb46-986b-4f0e-9f2b-6db3b006c975-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.648039 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.648016 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/1.log" Apr 21 14:27:42.648350 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.648335 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/0.log" Apr 21 14:27:42.648400 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.648370 2580 generic.go:358] "Generic (PLEG): container finished" podID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" containerID="983dff90097abb9b88de1e00cc8425fe86e06f9902a83f20cd76dda80a6fa3a9" exitCode=255 Apr 21 14:27:42.648452 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.648436 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" event={"ID":"1c4a98bb-ce18-4746-b5b2-979619d26ebc","Type":"ContainerDied","Data":"983dff90097abb9b88de1e00cc8425fe86e06f9902a83f20cd76dda80a6fa3a9"} Apr 21 14:27:42.648487 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.648475 2580 scope.go:117] "RemoveContainer" containerID="f810df9e65b4e91e7e06df2ffece6b708b9f8285659989129bfa60e47c54d6e6" Apr 21 14:27:42.648729 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.648696 2580 scope.go:117] "RemoveContainer" containerID="983dff90097abb9b88de1e00cc8425fe86e06f9902a83f20cd76dda80a6fa3a9" Apr 21 14:27:42.648919 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:42.648900 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-qjbtz_openshift-console-operator(1c4a98bb-ce18-4746-b5b2-979619d26ebc)\"" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" podUID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" Apr 21 14:27:42.710035 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.709995 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07d5eb46-986b-4f0e-9f2b-6db3b006c975-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.710199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.710060 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d69kz\" (UniqueName: \"kubernetes.io/projected/07d5eb46-986b-4f0e-9f2b-6db3b006c975-kube-api-access-d69kz\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.711291 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.710354 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07d5eb46-986b-4f0e-9f2b-6db3b006c975-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.711291 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.711052 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07d5eb46-986b-4f0e-9f2b-6db3b006c975-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.713971 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.713095 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07d5eb46-986b-4f0e-9f2b-6db3b006c975-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.719785 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.719765 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d69kz\" (UniqueName: \"kubernetes.io/projected/07d5eb46-986b-4f0e-9f2b-6db3b006c975-kube-api-access-d69kz\") pod \"kube-storage-version-migrator-operator-6769c5d45-2lrpd\" (UID: \"07d5eb46-986b-4f0e-9f2b-6db3b006c975\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.861383 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.861284 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" Apr 21 14:27:42.981448 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:42.981414 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd"] Apr 21 14:27:42.984486 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:27:42.984444 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07d5eb46_986b_4f0e_9f2b_6db3b006c975.slice/crio-f67d6dde373efd2f4c0ae1fabee1782a27a0591fe6fabecb75666a2551f52704 WatchSource:0}: Error finding container f67d6dde373efd2f4c0ae1fabee1782a27a0591fe6fabecb75666a2551f52704: Status 404 returned error can't find the container with id f67d6dde373efd2f4c0ae1fabee1782a27a0591fe6fabecb75666a2551f52704 Apr 21 14:27:43.317555 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.317501 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:43.317737 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.317691 2580 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:43.317798 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.317776 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls podName:f355a71a-6279-4fa0-9123-6a056ee88f72 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:51.317756682 +0000 UTC m=+104.775241191 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-pn2df" (UID: "f355a71a-6279-4fa0-9123-6a056ee88f72") : secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:43.418468 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.418419 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:43.418674 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.418527 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:43.418674 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.418590 2580 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 14:27:43.418674 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.418674 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:51.41865841 +0000 UTC m=+104.876142895 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : secret "router-metrics-certs-default" not found Apr 21 14:27:43.418785 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.418695 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:27:51.418680524 +0000 UTC m=+104.876165008 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : configmap references non-existent config key: service-ca.crt Apr 21 14:27:43.651389 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.651313 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/1.log" Apr 21 14:27:43.651796 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.651705 2580 scope.go:117] "RemoveContainer" containerID="983dff90097abb9b88de1e00cc8425fe86e06f9902a83f20cd76dda80a6fa3a9" Apr 21 14:27:43.651905 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.651877 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-qjbtz_openshift-console-operator(1c4a98bb-ce18-4746-b5b2-979619d26ebc)\"" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" podUID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" Apr 21 14:27:43.652411 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.652394 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" event={"ID":"07d5eb46-986b-4f0e-9f2b-6db3b006c975","Type":"ContainerStarted","Data":"f67d6dde373efd2f4c0ae1fabee1782a27a0591fe6fabecb75666a2551f52704"} Apr 21 14:27:43.822922 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.822886 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:27:43.822922 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:43.822927 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:27:43.823178 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.823034 2580 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 14:27:43.823178 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.823043 2580 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 14:27:43.823178 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.823082 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls podName:afea4c93-a952-4362-8775-179f3919eecc nodeName:}" failed. No retries permitted until 2026-04-21 14:28:47.823069299 +0000 UTC m=+161.280553784 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls") pod "dns-default-dx8mq" (UID: "afea4c93-a952-4362-8775-179f3919eecc") : secret "dns-default-metrics-tls" not found Apr 21 14:27:43.823178 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:43.823106 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert podName:5badb5d3-1012-4fbf-81cc-b5190659075e nodeName:}" failed. No retries permitted until 2026-04-21 14:28:47.823089588 +0000 UTC m=+161.280574088 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert") pod "ingress-canary-tfksq" (UID: "5badb5d3-1012-4fbf-81cc-b5190659075e") : secret "canary-serving-cert" not found Apr 21 14:27:45.335247 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.335223 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:45.335625 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:45.335388 2580 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 14:27:45.335625 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:45.335469 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls podName:4402dc96-86f0-45e6-b15c-7aace6e551be nodeName:}" failed. No retries permitted until 2026-04-21 14:27:53.335446729 +0000 UTC m=+106.792931215 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-fjvbh" (UID: "4402dc96-86f0-45e6-b15c-7aace6e551be") : secret "samples-operator-tls" not found Apr 21 14:27:45.658360 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.658263 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" event={"ID":"07d5eb46-986b-4f0e-9f2b-6db3b006c975","Type":"ContainerStarted","Data":"3be06cc8dd517999aafae4a0afc50c45546d1de11783e3b32a303ce2b52a5248"} Apr 21 14:27:45.675176 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.675116 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" podStartSLOduration=1.348796928 podStartE2EDuration="3.675100347s" podCreationTimestamp="2026-04-21 14:27:42 +0000 UTC" firstStartedPulling="2026-04-21 14:27:42.986420147 +0000 UTC m=+96.443904632" lastFinishedPulling="2026-04-21 14:27:45.312723563 +0000 UTC m=+98.770208051" observedRunningTime="2026-04-21 14:27:45.674041339 +0000 UTC m=+99.131525859" watchObservedRunningTime="2026-04-21 14:27:45.675100347 +0000 UTC m=+99.132584853" Apr 21 14:27:45.859916 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.859875 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j"] Apr 21 14:27:45.899699 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.899669 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j"] Apr 21 14:27:45.899880 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.899811 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" Apr 21 14:27:45.903366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.903342 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"network-diagnostics-dockercfg-ztp7s\"" Apr 21 14:27:45.939981 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:45.939951 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qk8n\" (UniqueName: \"kubernetes.io/projected/03d822e2-b682-4a3d-95fe-66f8e08b1585-kube-api-access-2qk8n\") pod \"network-check-source-8894fc9bd-jvt2j\" (UID: \"03d822e2-b682-4a3d-95fe-66f8e08b1585\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" Apr 21 14:27:46.041430 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.041391 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2qk8n\" (UniqueName: \"kubernetes.io/projected/03d822e2-b682-4a3d-95fe-66f8e08b1585-kube-api-access-2qk8n\") pod \"network-check-source-8894fc9bd-jvt2j\" (UID: \"03d822e2-b682-4a3d-95fe-66f8e08b1585\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" Apr 21 14:27:46.058045 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.057968 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qk8n\" (UniqueName: \"kubernetes.io/projected/03d822e2-b682-4a3d-95fe-66f8e08b1585-kube-api-access-2qk8n\") pod \"network-check-source-8894fc9bd-jvt2j\" (UID: \"03d822e2-b682-4a3d-95fe-66f8e08b1585\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" Apr 21 14:27:46.206615 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.206517 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-7c56c698b5-pb9q6"] Apr 21 14:27:46.222703 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.222675 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.227475 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.227434 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 21 14:27:46.227763 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.227741 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-4v5jp\"" Apr 21 14:27:46.229350 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.229324 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-7c56c698b5-pb9q6"] Apr 21 14:27:46.229915 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.229798 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 21 14:27:46.229915 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.229894 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 21 14:27:46.236696 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.236667 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 21 14:27:46.242280 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.242257 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" Apr 21 14:27:46.343238 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343204 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-trusted-ca\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.343653 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343260 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-certificates\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.343653 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343282 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-image-registry-private-configuration\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.343653 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343385 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-installation-pull-secrets\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.343653 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343474 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.343653 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343509 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-bound-sa-token\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.343653 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343555 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d85ed974-2064-4d7f-a10c-ff2a82315054-ca-trust-extracted\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.343653 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.343601 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44l2p\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-kube-api-access-44l2p\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.360894 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.360866 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j"] Apr 21 14:27:46.364100 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:27:46.364074 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03d822e2_b682_4a3d_95fe_66f8e08b1585.slice/crio-da0ddb2817574c1bb90b4405bfb95dde9d6beab42da8f14f3ddb6bf3cac67d95 WatchSource:0}: Error finding container da0ddb2817574c1bb90b4405bfb95dde9d6beab42da8f14f3ddb6bf3cac67d95: Status 404 returned error can't find the container with id da0ddb2817574c1bb90b4405bfb95dde9d6beab42da8f14f3ddb6bf3cac67d95 Apr 21 14:27:46.444785 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.444738 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445007 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.444795 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-bound-sa-token\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445007 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.444840 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d85ed974-2064-4d7f-a10c-ff2a82315054-ca-trust-extracted\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445007 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.444867 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-44l2p\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-kube-api-access-44l2p\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445007 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:46.444889 2580 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 14:27:46.445007 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:46.444908 2580 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7c56c698b5-pb9q6: secret "image-registry-tls" not found Apr 21 14:27:46.445007 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:46.444958 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls podName:d85ed974-2064-4d7f-a10c-ff2a82315054 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:46.944941499 +0000 UTC m=+100.402425983 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls") pod "image-registry-7c56c698b5-pb9q6" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054") : secret "image-registry-tls" not found Apr 21 14:27:46.445314 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.445035 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-trusted-ca\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445314 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.445076 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-certificates\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445314 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.445110 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-image-registry-private-configuration\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445314 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.445269 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d85ed974-2064-4d7f-a10c-ff2a82315054-ca-trust-extracted\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445512 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.445345 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-installation-pull-secrets\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.445727 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.445707 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-certificates\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.446008 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.445989 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-trusted-ca\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.448218 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.448168 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-installation-pull-secrets\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.448323 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.448219 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-image-registry-private-configuration\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.457674 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.457601 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-44l2p\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-kube-api-access-44l2p\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.459154 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.459132 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-bound-sa-token\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.587758 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.587724 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-j6wgt" Apr 21 14:27:46.662556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.662520 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" event={"ID":"03d822e2-b682-4a3d-95fe-66f8e08b1585","Type":"ContainerStarted","Data":"983805326b0b89f214284d093e45ceb8831be65ca84da03f42fa79b9a8299f82"} Apr 21 14:27:46.662556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.662557 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" event={"ID":"03d822e2-b682-4a3d-95fe-66f8e08b1585","Type":"ContainerStarted","Data":"da0ddb2817574c1bb90b4405bfb95dde9d6beab42da8f14f3ddb6bf3cac67d95"} Apr 21 14:27:46.682668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.682619 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-jvt2j" podStartSLOduration=1.682605452 podStartE2EDuration="1.682605452s" podCreationTimestamp="2026-04-21 14:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:27:46.682097898 +0000 UTC m=+100.139582406" watchObservedRunningTime="2026-04-21 14:27:46.682605452 +0000 UTC m=+100.140089959" Apr 21 14:27:46.950965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:46.950916 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:46.951176 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:46.951084 2580 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 14:27:46.951176 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:46.951108 2580 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7c56c698b5-pb9q6: secret "image-registry-tls" not found Apr 21 14:27:46.951299 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:46.951194 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls podName:d85ed974-2064-4d7f-a10c-ff2a82315054 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:47.951171994 +0000 UTC m=+101.408656494 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls") pod "image-registry-7c56c698b5-pb9q6" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054") : secret "image-registry-tls" not found Apr 21 14:27:47.960406 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:47.960367 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:47.960798 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:47.960523 2580 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 14:27:47.960798 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:47.960544 2580 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7c56c698b5-pb9q6: secret "image-registry-tls" not found Apr 21 14:27:47.960798 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:47.960623 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls podName:d85ed974-2064-4d7f-a10c-ff2a82315054 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:49.960608425 +0000 UTC m=+103.418092910 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls") pod "image-registry-7c56c698b5-pb9q6" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054") : secret "image-registry-tls" not found Apr 21 14:27:48.824687 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:48.824636 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:48.824687 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:48.824693 2580 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:27:48.825058 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:48.825046 2580 scope.go:117] "RemoveContainer" containerID="983dff90097abb9b88de1e00cc8425fe86e06f9902a83f20cd76dda80a6fa3a9" Apr 21 14:27:48.825216 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:48.825201 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-qjbtz_openshift-console-operator(1c4a98bb-ce18-4746-b5b2-979619d26ebc)\"" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" podUID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" Apr 21 14:27:49.978886 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:49.978852 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:49.979341 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:49.979031 2580 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 14:27:49.979341 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:49.979057 2580 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7c56c698b5-pb9q6: secret "image-registry-tls" not found Apr 21 14:27:49.979341 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:49.979130 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls podName:d85ed974-2064-4d7f-a10c-ff2a82315054 nodeName:}" failed. No retries permitted until 2026-04-21 14:27:53.979108356 +0000 UTC m=+107.436592841 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls") pod "image-registry-7c56c698b5-pb9q6" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054") : secret "image-registry-tls" not found Apr 21 14:27:51.388988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.388943 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:27:51.389373 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:51.389069 2580 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:51.389373 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:51.389124 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls podName:f355a71a-6279-4fa0-9123-6a056ee88f72 nodeName:}" failed. No retries permitted until 2026-04-21 14:28:07.389110668 +0000 UTC m=+120.846595152 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-pn2df" (UID: "f355a71a-6279-4fa0-9123-6a056ee88f72") : secret "cluster-monitoring-operator-tls" not found Apr 21 14:27:51.489986 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.489951 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:51.490138 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.490039 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:27:51.490138 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:51.490095 2580 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 14:27:51.490205 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:51.490154 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:28:07.490141399 +0000 UTC m=+120.947625883 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : configmap references non-existent config key: service-ca.crt Apr 21 14:27:51.490205 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:51.490168 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs podName:0284ee69-97a8-4c33-98ee-1bb75e7f461b nodeName:}" failed. No retries permitted until 2026-04-21 14:28:07.49016267 +0000 UTC m=+120.947647154 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs") pod "router-default-6d6c54d4dd-mb7h8" (UID: "0284ee69-97a8-4c33-98ee-1bb75e7f461b") : secret "router-metrics-certs-default" not found Apr 21 14:27:51.872699 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.872662 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-865cb79987-65rj2"] Apr 21 14:27:51.875401 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.875383 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:51.878265 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.878242 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Apr 21 14:27:51.878855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.878835 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Apr 21 14:27:51.879218 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.879203 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-4qfs5\"" Apr 21 14:27:51.879218 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.879212 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Apr 21 14:27:51.879310 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.879249 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Apr 21 14:27:51.884973 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.884947 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-65rj2"] Apr 21 14:27:51.994300 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.994260 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbvz2\" (UniqueName: \"kubernetes.io/projected/3eaef19d-8a64-4260-bb7a-baf55a9694b1-kube-api-access-rbvz2\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:51.994462 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.994313 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3eaef19d-8a64-4260-bb7a-baf55a9694b1-signing-cabundle\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:51.994462 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:51.994415 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3eaef19d-8a64-4260-bb7a-baf55a9694b1-signing-key\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.095072 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.095022 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3eaef19d-8a64-4260-bb7a-baf55a9694b1-signing-cabundle\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.095189 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.095111 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3eaef19d-8a64-4260-bb7a-baf55a9694b1-signing-key\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.095189 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.095182 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rbvz2\" (UniqueName: \"kubernetes.io/projected/3eaef19d-8a64-4260-bb7a-baf55a9694b1-kube-api-access-rbvz2\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.095868 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.095797 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3eaef19d-8a64-4260-bb7a-baf55a9694b1-signing-cabundle\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.097675 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.097656 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3eaef19d-8a64-4260-bb7a-baf55a9694b1-signing-key\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.103910 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.103889 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbvz2\" (UniqueName: \"kubernetes.io/projected/3eaef19d-8a64-4260-bb7a-baf55a9694b1-kube-api-access-rbvz2\") pod \"service-ca-865cb79987-65rj2\" (UID: \"3eaef19d-8a64-4260-bb7a-baf55a9694b1\") " pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.185646 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.185619 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-65rj2" Apr 21 14:27:52.302034 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.302008 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-65rj2"] Apr 21 14:27:52.304959 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:27:52.304934 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3eaef19d_8a64_4260_bb7a_baf55a9694b1.slice/crio-df80d9eb1f264e1d489b2baa3f3087f969a140f2db395a52ba131d33f1703606 WatchSource:0}: Error finding container df80d9eb1f264e1d489b2baa3f3087f969a140f2db395a52ba131d33f1703606: Status 404 returned error can't find the container with id df80d9eb1f264e1d489b2baa3f3087f969a140f2db395a52ba131d33f1703606 Apr 21 14:27:52.681447 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:52.681413 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-65rj2" event={"ID":"3eaef19d-8a64-4260-bb7a-baf55a9694b1","Type":"ContainerStarted","Data":"df80d9eb1f264e1d489b2baa3f3087f969a140f2db395a52ba131d33f1703606"} Apr 21 14:27:53.407234 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:53.407183 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:27:53.407414 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:53.407338 2580 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 14:27:53.407414 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:53.407406 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls podName:4402dc96-86f0-45e6-b15c-7aace6e551be nodeName:}" failed. No retries permitted until 2026-04-21 14:28:09.407389267 +0000 UTC m=+122.864873752 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-fjvbh" (UID: "4402dc96-86f0-45e6-b15c-7aace6e551be") : secret "samples-operator-tls" not found Apr 21 14:27:54.013735 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:54.011928 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:27:54.013735 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:54.012079 2580 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 14:27:54.013735 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:54.012096 2580 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7c56c698b5-pb9q6: secret "image-registry-tls" not found Apr 21 14:27:54.013735 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:27:54.012161 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls podName:d85ed974-2064-4d7f-a10c-ff2a82315054 nodeName:}" failed. No retries permitted until 2026-04-21 14:28:02.012142076 +0000 UTC m=+115.469626580 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls") pod "image-registry-7c56c698b5-pb9q6" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054") : secret "image-registry-tls" not found Apr 21 14:27:55.691045 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:55.691011 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-65rj2" event={"ID":"3eaef19d-8a64-4260-bb7a-baf55a9694b1","Type":"ContainerStarted","Data":"db813f8b18a470f8108a55c06b61125ca5677c975cf8b19eba65083a7240e94e"} Apr 21 14:27:55.706990 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:27:55.706923 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-865cb79987-65rj2" podStartSLOduration=1.903635038 podStartE2EDuration="4.706910296s" podCreationTimestamp="2026-04-21 14:27:51 +0000 UTC" firstStartedPulling="2026-04-21 14:27:52.307180527 +0000 UTC m=+105.764665013" lastFinishedPulling="2026-04-21 14:27:55.110455783 +0000 UTC m=+108.567940271" observedRunningTime="2026-04-21 14:27:55.70643849 +0000 UTC m=+109.163922997" watchObservedRunningTime="2026-04-21 14:27:55.706910296 +0000 UTC m=+109.164394803" Apr 21 14:28:00.242010 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:00.241971 2580 scope.go:117] "RemoveContainer" containerID="983dff90097abb9b88de1e00cc8425fe86e06f9902a83f20cd76dda80a6fa3a9" Apr 21 14:28:00.705474 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:00.705446 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:28:00.705855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:00.705840 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/1.log" Apr 21 14:28:00.705930 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:00.705873 2580 generic.go:358] "Generic (PLEG): container finished" podID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" containerID="14234e58bc12735d2a941d445ad449f1db0cc2b040f80ae0bbeb97b8d71be5ba" exitCode=255 Apr 21 14:28:00.705985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:00.705933 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" event={"ID":"1c4a98bb-ce18-4746-b5b2-979619d26ebc","Type":"ContainerDied","Data":"14234e58bc12735d2a941d445ad449f1db0cc2b040f80ae0bbeb97b8d71be5ba"} Apr 21 14:28:00.705985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:00.705963 2580 scope.go:117] "RemoveContainer" containerID="983dff90097abb9b88de1e00cc8425fe86e06f9902a83f20cd76dda80a6fa3a9" Apr 21 14:28:00.706371 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:00.706351 2580 scope.go:117] "RemoveContainer" containerID="14234e58bc12735d2a941d445ad449f1db0cc2b040f80ae0bbeb97b8d71be5ba" Apr 21 14:28:00.706586 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:28:00.706541 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-qjbtz_openshift-console-operator(1c4a98bb-ce18-4746-b5b2-979619d26ebc)\"" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" podUID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" Apr 21 14:28:01.710115 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:01.710083 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:28:02.080810 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.080710 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:28:02.083264 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.083238 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"image-registry-7c56c698b5-pb9q6\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:28:02.134835 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.134805 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:28:02.265913 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.265875 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-7c56c698b5-pb9q6"] Apr 21 14:28:02.269866 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:02.269840 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd85ed974_2064_4d7f_a10c_ff2a82315054.slice/crio-ab6aa87fc09aeec322b58e19e6d2180b3ab0f1c6f670496e89e39e0b18648db2 WatchSource:0}: Error finding container ab6aa87fc09aeec322b58e19e6d2180b3ab0f1c6f670496e89e39e0b18648db2: Status 404 returned error can't find the container with id ab6aa87fc09aeec322b58e19e6d2180b3ab0f1c6f670496e89e39e0b18648db2 Apr 21 14:28:02.714547 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.714513 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" event={"ID":"d85ed974-2064-4d7f-a10c-ff2a82315054","Type":"ContainerStarted","Data":"6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c"} Apr 21 14:28:02.714985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.714556 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" event={"ID":"d85ed974-2064-4d7f-a10c-ff2a82315054","Type":"ContainerStarted","Data":"ab6aa87fc09aeec322b58e19e6d2180b3ab0f1c6f670496e89e39e0b18648db2"} Apr 21 14:28:02.714985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.714690 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:28:02.733033 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:02.732965 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" podStartSLOduration=16.732945209 podStartE2EDuration="16.732945209s" podCreationTimestamp="2026-04-21 14:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:28:02.732368674 +0000 UTC m=+116.189853205" watchObservedRunningTime="2026-04-21 14:28:02.732945209 +0000 UTC m=+116.190429721" Apr 21 14:28:07.427852 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.427816 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:28:07.430303 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.430278 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/f355a71a-6279-4fa0-9123-6a056ee88f72-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-pn2df\" (UID: \"f355a71a-6279-4fa0-9123-6a056ee88f72\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:28:07.441307 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.441275 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-r7pvz\"" Apr 21 14:28:07.450188 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.450164 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" Apr 21 14:28:07.529194 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.529053 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:07.529194 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.529174 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:07.529986 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.529965 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0284ee69-97a8-4c33-98ee-1bb75e7f461b-service-ca-bundle\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:07.537704 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.537675 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0284ee69-97a8-4c33-98ee-1bb75e7f461b-metrics-certs\") pod \"router-default-6d6c54d4dd-mb7h8\" (UID: \"0284ee69-97a8-4c33-98ee-1bb75e7f461b\") " pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:07.541302 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.541274 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-x64rj\"" Apr 21 14:28:07.549654 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.549633 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:07.576542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.576518 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df"] Apr 21 14:28:07.579346 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:07.579315 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf355a71a_6279_4fa0_9123_6a056ee88f72.slice/crio-94103b45f08f1a4cd22cae7e5821fcb2eeba74249d05c86975d166668f00619d WatchSource:0}: Error finding container 94103b45f08f1a4cd22cae7e5821fcb2eeba74249d05c86975d166668f00619d: Status 404 returned error can't find the container with id 94103b45f08f1a4cd22cae7e5821fcb2eeba74249d05c86975d166668f00619d Apr 21 14:28:07.676683 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.676656 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/router-default-6d6c54d4dd-mb7h8"] Apr 21 14:28:07.679049 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:07.679000 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0284ee69_97a8_4c33_98ee_1bb75e7f461b.slice/crio-38c03faf501e36963a3cbb97ef981ab88cd8eb05edc7750c91bf35a66e8e21ec WatchSource:0}: Error finding container 38c03faf501e36963a3cbb97ef981ab88cd8eb05edc7750c91bf35a66e8e21ec: Status 404 returned error can't find the container with id 38c03faf501e36963a3cbb97ef981ab88cd8eb05edc7750c91bf35a66e8e21ec Apr 21 14:28:07.728630 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.728599 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" event={"ID":"0284ee69-97a8-4c33-98ee-1bb75e7f461b","Type":"ContainerStarted","Data":"38c03faf501e36963a3cbb97ef981ab88cd8eb05edc7750c91bf35a66e8e21ec"} Apr 21 14:28:07.729444 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:07.729421 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" event={"ID":"f355a71a-6279-4fa0-9123-6a056ee88f72","Type":"ContainerStarted","Data":"94103b45f08f1a4cd22cae7e5821fcb2eeba74249d05c86975d166668f00619d"} Apr 21 14:28:08.733315 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:08.733278 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" event={"ID":"0284ee69-97a8-4c33-98ee-1bb75e7f461b","Type":"ContainerStarted","Data":"0b0f54630eb1171f0d9ca7b8a709eb6b62d8b50514b0d97688c6054c459ec71a"} Apr 21 14:28:08.753966 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:08.753921 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" podStartSLOduration=33.753905735000004 podStartE2EDuration="33.753905735s" podCreationTimestamp="2026-04-21 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:28:08.753786745 +0000 UTC m=+122.211271251" watchObservedRunningTime="2026-04-21 14:28:08.753905735 +0000 UTC m=+122.211390297" Apr 21 14:28:08.824247 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:08.824208 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:28:08.824430 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:08.824261 2580 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:28:08.824784 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:08.824748 2580 scope.go:117] "RemoveContainer" containerID="14234e58bc12735d2a941d445ad449f1db0cc2b040f80ae0bbeb97b8d71be5ba" Apr 21 14:28:08.825040 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:28:08.825011 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-qjbtz_openshift-console-operator(1c4a98bb-ce18-4746-b5b2-979619d26ebc)\"" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" podUID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" Apr 21 14:28:09.448225 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.448177 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:28:09.450998 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.450969 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4402dc96-86f0-45e6-b15c-7aace6e551be-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-fjvbh\" (UID: \"4402dc96-86f0-45e6-b15c-7aace6e551be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:28:09.455888 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.455862 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-zx6fl\"" Apr 21 14:28:09.464715 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.464689 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" Apr 21 14:28:09.550242 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.550214 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:09.552835 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.552803 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:09.607783 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.607760 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh"] Apr 21 14:28:09.737216 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.737177 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" event={"ID":"4402dc96-86f0-45e6-b15c-7aace6e551be","Type":"ContainerStarted","Data":"b7123c9b64347cd6982141b9b5d6a99915d6aea2a29b3aa6328cf15947661446"} Apr 21 14:28:09.738616 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.738589 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" event={"ID":"f355a71a-6279-4fa0-9123-6a056ee88f72","Type":"ContainerStarted","Data":"ffd6c44893184d4d13c4696105e1e8aefa35ea88ed72d87494e6eadb4e2f6ad0"} Apr 21 14:28:09.739053 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.738868 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:09.740169 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.740150 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-6d6c54d4dd-mb7h8" Apr 21 14:28:09.756409 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:09.756362 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" podStartSLOduration=33.087521729 podStartE2EDuration="34.756346343s" podCreationTimestamp="2026-04-21 14:27:35 +0000 UTC" firstStartedPulling="2026-04-21 14:28:07.58117035 +0000 UTC m=+121.038654835" lastFinishedPulling="2026-04-21 14:28:09.249994964 +0000 UTC m=+122.707479449" observedRunningTime="2026-04-21 14:28:09.754879753 +0000 UTC m=+123.212364259" watchObservedRunningTime="2026-04-21 14:28:09.756346343 +0000 UTC m=+123.213830854" Apr 21 14:28:12.456275 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.456240 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-7c56c698b5-pb9q6"] Apr 21 14:28:12.497329 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.497289 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-79f755579b-8wxj4"] Apr 21 14:28:12.528596 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.528550 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-79f755579b-8wxj4"] Apr 21 14:28:12.528770 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.528723 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.576998 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.576968 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-ncz6v"] Apr 21 14:28:12.595523 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.595493 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-ncz6v"] Apr 21 14:28:12.595703 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.595635 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.597753 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.597729 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-qlm9d\"" Apr 21 14:28:12.597904 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.597824 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 21 14:28:12.597904 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.597828 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 21 14:28:12.679767 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679733 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-registry-tls\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.679767 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679776 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-installation-pull-secrets\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.679982 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679843 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-trusted-ca\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.679982 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679872 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-bound-sa-token\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.679982 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679893 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7g46\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-kube-api-access-d7g46\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.679982 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679923 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-ca-trust-extracted\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.679982 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679946 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-registry-certificates\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.680134 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.679986 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-image-registry-private-configuration\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.748891 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.748810 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" event={"ID":"4402dc96-86f0-45e6-b15c-7aace6e551be","Type":"ContainerStarted","Data":"26c819a360a46f371c2a2aad0bdbf686d46a2f379ed45c8b13ddb2dd8ba28926"} Apr 21 14:28:12.748891 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.748847 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" event={"ID":"4402dc96-86f0-45e6-b15c-7aace6e551be","Type":"ContainerStarted","Data":"c095f06570e37d85c69809ac9db82c1e1cccf2805dc259d8f95b93df1649a56f"} Apr 21 14:28:12.770464 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.770412 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-fjvbh" podStartSLOduration=33.700685213 podStartE2EDuration="35.770397018s" podCreationTimestamp="2026-04-21 14:27:37 +0000 UTC" firstStartedPulling="2026-04-21 14:28:09.655215349 +0000 UTC m=+123.112699838" lastFinishedPulling="2026-04-21 14:28:11.724927145 +0000 UTC m=+125.182411643" observedRunningTime="2026-04-21 14:28:12.769325956 +0000 UTC m=+126.226810456" watchObservedRunningTime="2026-04-21 14:28:12.770397018 +0000 UTC m=+126.227881643" Apr 21 14:28:12.781252 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781224 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-image-registry-private-configuration\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781438 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781271 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/95687f91-0372-4f50-b65d-1c77acceeaeb-crio-socket\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.781438 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781314 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7pgh\" (UniqueName: \"kubernetes.io/projected/95687f91-0372-4f50-b65d-1c77acceeaeb-kube-api-access-v7pgh\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.781438 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781364 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/95687f91-0372-4f50-b65d-1c77acceeaeb-data-volume\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.781438 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781408 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/95687f91-0372-4f50-b65d-1c77acceeaeb-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.781672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781466 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-registry-tls\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781497 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-installation-pull-secrets\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781531 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-trusted-ca\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781557 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-bound-sa-token\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781610 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d7g46\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-kube-api-access-d7g46\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781657 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-ca-trust-extracted\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781963 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781845 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-registry-certificates\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.781963 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.781887 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/95687f91-0372-4f50-b65d-1c77acceeaeb-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.782134 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.782115 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-ca-trust-extracted\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.782545 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.782523 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-trusted-ca\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.782665 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.782629 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-registry-certificates\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.784025 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.784008 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-installation-pull-secrets\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.784427 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.784413 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-image-registry-private-configuration\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.784660 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.784641 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-registry-tls\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.796058 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.796028 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-bound-sa-token\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.796192 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.796175 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7g46\" (UniqueName: \"kubernetes.io/projected/0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5-kube-api-access-d7g46\") pod \"image-registry-79f755579b-8wxj4\" (UID: \"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5\") " pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.842211 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.842178 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:12.882905 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.882857 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v7pgh\" (UniqueName: \"kubernetes.io/projected/95687f91-0372-4f50-b65d-1c77acceeaeb-kube-api-access-v7pgh\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.883087 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.882922 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/95687f91-0372-4f50-b65d-1c77acceeaeb-data-volume\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.883087 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.882952 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/95687f91-0372-4f50-b65d-1c77acceeaeb-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.883087 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.883031 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/95687f91-0372-4f50-b65d-1c77acceeaeb-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.883330 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.883301 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/95687f91-0372-4f50-b65d-1c77acceeaeb-data-volume\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.883330 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.883305 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/95687f91-0372-4f50-b65d-1c77acceeaeb-crio-socket\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.883524 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.883353 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/95687f91-0372-4f50-b65d-1c77acceeaeb-crio-socket\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.883725 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.883708 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/95687f91-0372-4f50-b65d-1c77acceeaeb-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.885722 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.885701 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/95687f91-0372-4f50-b65d-1c77acceeaeb-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.892399 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.892373 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7pgh\" (UniqueName: \"kubernetes.io/projected/95687f91-0372-4f50-b65d-1c77acceeaeb-kube-api-access-v7pgh\") pod \"insights-runtime-extractor-ncz6v\" (UID: \"95687f91-0372-4f50-b65d-1c77acceeaeb\") " pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.905649 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.905615 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-ncz6v" Apr 21 14:28:12.976961 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:12.976932 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-79f755579b-8wxj4"] Apr 21 14:28:12.980352 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:12.980324 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d04ceb9_3ece_4a49_997b_3c8a3e6fc1f5.slice/crio-fee8343b161847f539cda3c7e9ba80d9ce1a0f7657ae0b53fb84f6ebb158d82f WatchSource:0}: Error finding container fee8343b161847f539cda3c7e9ba80d9ce1a0f7657ae0b53fb84f6ebb158d82f: Status 404 returned error can't find the container with id fee8343b161847f539cda3c7e9ba80d9ce1a0f7657ae0b53fb84f6ebb158d82f Apr 21 14:28:13.027029 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:13.026975 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-ncz6v"] Apr 21 14:28:13.035092 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:13.035063 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95687f91_0372_4f50_b65d_1c77acceeaeb.slice/crio-0b2bfb9d6483cc646623cb1c6a62791f17a7d3f7566de4e1b6bdc66689e59e1e WatchSource:0}: Error finding container 0b2bfb9d6483cc646623cb1c6a62791f17a7d3f7566de4e1b6bdc66689e59e1e: Status 404 returned error can't find the container with id 0b2bfb9d6483cc646623cb1c6a62791f17a7d3f7566de4e1b6bdc66689e59e1e Apr 21 14:28:13.752814 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:13.752779 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-79f755579b-8wxj4" event={"ID":"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5","Type":"ContainerStarted","Data":"da0070de8abe25c91bc235f39d61ab2a58b393af270d96563c0bd2d7636593fa"} Apr 21 14:28:13.752814 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:13.752817 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-79f755579b-8wxj4" event={"ID":"0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5","Type":"ContainerStarted","Data":"fee8343b161847f539cda3c7e9ba80d9ce1a0f7657ae0b53fb84f6ebb158d82f"} Apr 21 14:28:13.753432 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:13.752862 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:13.754142 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:13.754115 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ncz6v" event={"ID":"95687f91-0372-4f50-b65d-1c77acceeaeb","Type":"ContainerStarted","Data":"3f00d7ea18935419b1e91e20e5c07a9239274c35dc924b4e7ed25ed3ff7c9157"} Apr 21 14:28:13.754142 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:13.754144 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ncz6v" event={"ID":"95687f91-0372-4f50-b65d-1c77acceeaeb","Type":"ContainerStarted","Data":"0b2bfb9d6483cc646623cb1c6a62791f17a7d3f7566de4e1b6bdc66689e59e1e"} Apr 21 14:28:13.775329 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:13.775269 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-79f755579b-8wxj4" podStartSLOduration=1.775251918 podStartE2EDuration="1.775251918s" podCreationTimestamp="2026-04-21 14:28:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:28:13.772334431 +0000 UTC m=+127.229818939" watchObservedRunningTime="2026-04-21 14:28:13.775251918 +0000 UTC m=+127.232736426" Apr 21 14:28:14.758867 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:14.758834 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ncz6v" event={"ID":"95687f91-0372-4f50-b65d-1c77acceeaeb","Type":"ContainerStarted","Data":"7638c50f22009b244cd283a73ddb10e684fd78723ffa6c471d19dadac581dc6e"} Apr 21 14:28:15.764387 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:15.764351 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ncz6v" event={"ID":"95687f91-0372-4f50-b65d-1c77acceeaeb","Type":"ContainerStarted","Data":"04a3542b399bdfb263c89bae385409c4de012fbd89ee54ed27801b1b1ce0c2b6"} Apr 21 14:28:15.782167 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:15.782122 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-ncz6v" podStartSLOduration=1.313636242 podStartE2EDuration="3.78210757s" podCreationTimestamp="2026-04-21 14:28:12 +0000 UTC" firstStartedPulling="2026-04-21 14:28:13.111879432 +0000 UTC m=+126.569363920" lastFinishedPulling="2026-04-21 14:28:15.58035076 +0000 UTC m=+129.037835248" observedRunningTime="2026-04-21 14:28:15.781210669 +0000 UTC m=+129.238695175" watchObservedRunningTime="2026-04-21 14:28:15.78210757 +0000 UTC m=+129.239592078" Apr 21 14:28:15.910266 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:15.910175 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:28:15.912588 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:15.912547 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f89a533-2e7f-4655-bd1f-cf6e02cc46ed-metrics-certs\") pod \"network-metrics-daemon-fl65d\" (UID: \"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed\") " pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:28:16.056229 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:16.056191 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-9wlmf\"" Apr 21 14:28:16.064258 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:16.064234 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fl65d" Apr 21 14:28:16.187447 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:16.187421 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fl65d"] Apr 21 14:28:16.189978 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:16.189949 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f89a533_2e7f_4655_bd1f_cf6e02cc46ed.slice/crio-4d1bbeacd95024e87196423cc2c9409480964d00f15440bb53335a58b5f56f87 WatchSource:0}: Error finding container 4d1bbeacd95024e87196423cc2c9409480964d00f15440bb53335a58b5f56f87: Status 404 returned error can't find the container with id 4d1bbeacd95024e87196423cc2c9409480964d00f15440bb53335a58b5f56f87 Apr 21 14:28:16.768095 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:16.768052 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fl65d" event={"ID":"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed","Type":"ContainerStarted","Data":"4d1bbeacd95024e87196423cc2c9409480964d00f15440bb53335a58b5f56f87"} Apr 21 14:28:17.771868 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:17.771831 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fl65d" event={"ID":"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed","Type":"ContainerStarted","Data":"2fea27d9c5933a7c3ce4783ca42519ac3adf1c012e2479d48899ce105cada60c"} Apr 21 14:28:17.771868 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:17.771873 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fl65d" event={"ID":"2f89a533-2e7f-4655-bd1f-cf6e02cc46ed","Type":"ContainerStarted","Data":"7d96de93ef81741bbc277886327e45975f9925df3777e45efa9c0a1df30cb6a2"} Apr 21 14:28:17.788728 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:17.788679 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-fl65d" podStartSLOduration=129.66492997 podStartE2EDuration="2m10.788666557s" podCreationTimestamp="2026-04-21 14:26:07 +0000 UTC" firstStartedPulling="2026-04-21 14:28:16.192163381 +0000 UTC m=+129.649647866" lastFinishedPulling="2026-04-21 14:28:17.315899949 +0000 UTC m=+130.773384453" observedRunningTime="2026-04-21 14:28:17.787904858 +0000 UTC m=+131.245389366" watchObservedRunningTime="2026-04-21 14:28:17.788666557 +0000 UTC m=+131.246151109" Apr 21 14:28:19.238358 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.238309 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-ndjgn"] Apr 21 14:28:19.241531 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.241509 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.244290 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.244040 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 21 14:28:19.244290 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.244061 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 21 14:28:19.244290 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.244287 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 21 14:28:19.244675 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.244604 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 21 14:28:19.244675 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.244645 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-k54lk\"" Apr 21 14:28:19.334645 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334606 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-textfile\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.334834 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334687 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-metrics-client-ca\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.334834 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334718 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-accelerators-collector-config\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.334834 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334743 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-885zj\" (UniqueName: \"kubernetes.io/projected/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-kube-api-access-885zj\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.334834 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334773 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-root\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.335011 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334831 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-sys\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.335011 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334863 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.335011 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334922 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-tls\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.335011 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.334997 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-wtmp\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.435585 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.435536 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-metrics-client-ca\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.435767 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.435601 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-accelerators-collector-config\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.435767 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.435629 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-885zj\" (UniqueName: \"kubernetes.io/projected/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-kube-api-access-885zj\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.435767 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.435660 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-root\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.435767 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.435686 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-sys\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.435767 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.435715 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436189 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-root\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436226 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-tls\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436284 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-accelerators-collector-config\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436305 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-wtmp\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:28:19.436338 2580 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436377 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-textfile\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436434 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-wtmp\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.436652 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:28:19.436408 2580 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-tls podName:48f962b2-2b0a-455d-91f8-1a90cdb8f2db nodeName:}" failed. No retries permitted until 2026-04-21 14:28:19.936388004 +0000 UTC m=+133.393872492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-tls") pod "node-exporter-ndjgn" (UID: "48f962b2-2b0a-455d-91f8-1a90cdb8f2db") : secret "node-exporter-tls" not found Apr 21 14:28:19.437137 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436725 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-textfile\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.437137 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.436819 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-metrics-client-ca\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.437585 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.437544 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-sys\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.438337 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.438317 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.446252 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.446231 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-885zj\" (UniqueName: \"kubernetes.io/projected/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-kube-api-access-885zj\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.941198 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.941165 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-tls\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:19.944002 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:19.943974 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/48f962b2-2b0a-455d-91f8-1a90cdb8f2db-node-exporter-tls\") pod \"node-exporter-ndjgn\" (UID: \"48f962b2-2b0a-455d-91f8-1a90cdb8f2db\") " pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:20.153103 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:20.153076 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-ndjgn" Apr 21 14:28:20.162027 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:20.161994 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48f962b2_2b0a_455d_91f8_1a90cdb8f2db.slice/crio-a9821436aaeff127eaadd1ae2dfc08401f490ea48284925947144724b5fd8e80 WatchSource:0}: Error finding container a9821436aaeff127eaadd1ae2dfc08401f490ea48284925947144724b5fd8e80: Status 404 returned error can't find the container with id a9821436aaeff127eaadd1ae2dfc08401f490ea48284925947144724b5fd8e80 Apr 21 14:28:20.781029 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:20.780993 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-ndjgn" event={"ID":"48f962b2-2b0a-455d-91f8-1a90cdb8f2db","Type":"ContainerStarted","Data":"a9821436aaeff127eaadd1ae2dfc08401f490ea48284925947144724b5fd8e80"} Apr 21 14:28:21.241276 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:21.241245 2580 scope.go:117] "RemoveContainer" containerID="14234e58bc12735d2a941d445ad449f1db0cc2b040f80ae0bbeb97b8d71be5ba" Apr 21 14:28:21.788502 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:21.788424 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:28:21.788952 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:21.788543 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" event={"ID":"1c4a98bb-ce18-4746-b5b2-979619d26ebc","Type":"ContainerStarted","Data":"ae10220a969b3ab02c0b2e54b6d33028526c94489512dd6dd22fb2deb7b11a85"} Apr 21 14:28:21.788952 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:21.788863 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:28:21.790262 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:21.790234 2580 generic.go:358] "Generic (PLEG): container finished" podID="48f962b2-2b0a-455d-91f8-1a90cdb8f2db" containerID="1ba035e5cebb91199bbd7d8fdbb55283638616aa5cceb3f65af63d10d2f016e8" exitCode=0 Apr 21 14:28:21.790405 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:21.790271 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-ndjgn" event={"ID":"48f962b2-2b0a-455d-91f8-1a90cdb8f2db","Type":"ContainerDied","Data":"1ba035e5cebb91199bbd7d8fdbb55283638616aa5cceb3f65af63d10d2f016e8"} Apr 21 14:28:21.806263 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:21.806193 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" podStartSLOduration=42.076399099 podStartE2EDuration="43.806177875s" podCreationTimestamp="2026-04-21 14:27:38 +0000 UTC" firstStartedPulling="2026-04-21 14:27:39.242520686 +0000 UTC m=+92.700005175" lastFinishedPulling="2026-04-21 14:27:40.972299466 +0000 UTC m=+94.429783951" observedRunningTime="2026-04-21 14:28:21.804551821 +0000 UTC m=+135.262036327" watchObservedRunningTime="2026-04-21 14:28:21.806177875 +0000 UTC m=+135.263662386" Apr 21 14:28:22.461727 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:22.461693 2580 patch_prober.go:28] interesting pod/image-registry-7c56c698b5-pb9q6 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 21 14:28:22.461912 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:22.461754 2580 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" podUID="d85ed974-2064-4d7f-a10c-ff2a82315054" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 21 14:28:22.789812 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:22.789721 2580 patch_prober.go:28] interesting pod/console-operator-9d4b6777b-qjbtz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.134.0.12:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Apr 21 14:28:22.789812 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:22.789783 2580 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" podUID="1c4a98bb-ce18-4746-b5b2-979619d26ebc" containerName="console-operator" probeResult="failure" output="Get \"https://10.134.0.12:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Apr 21 14:28:22.795165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:22.795126 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-ndjgn" event={"ID":"48f962b2-2b0a-455d-91f8-1a90cdb8f2db","Type":"ContainerStarted","Data":"3ac5d0101a3631085613a6b5b5abf9b588b72a4b696371807ff4622a18bc98fa"} Apr 21 14:28:22.795321 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:22.795170 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-ndjgn" event={"ID":"48f962b2-2b0a-455d-91f8-1a90cdb8f2db","Type":"ContainerStarted","Data":"28df6ddd29f9c0674649b56f963d15c0f7ee3d94a2540de4c7253b2ca524c0de"} Apr 21 14:28:22.814294 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:22.814236 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-ndjgn" podStartSLOduration=2.669385976 podStartE2EDuration="3.814219703s" podCreationTimestamp="2026-04-21 14:28:19 +0000 UTC" firstStartedPulling="2026-04-21 14:28:20.163806967 +0000 UTC m=+133.621291457" lastFinishedPulling="2026-04-21 14:28:21.308640697 +0000 UTC m=+134.766125184" observedRunningTime="2026-04-21 14:28:22.813879105 +0000 UTC m=+136.271363622" watchObservedRunningTime="2026-04-21 14:28:22.814219703 +0000 UTC m=+136.271704212" Apr 21 14:28:23.074775 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:23.074699 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-9d4b6777b-qjbtz" Apr 21 14:28:23.984772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:23.984739 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh"] Apr 21 14:28:23.987190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:23.987173 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" Apr 21 14:28:23.991330 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:23.991307 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"default-dockercfg-l5wcf\"" Apr 21 14:28:23.991462 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:23.991312 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"monitoring-plugin-cert\"" Apr 21 14:28:23.999143 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:23.999118 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh"] Apr 21 14:28:24.079212 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:24.079179 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1779e5fc-59fd-4957-a145-e08e128869c7-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-fk6qh\" (UID: \"1779e5fc-59fd-4957-a145-e08e128869c7\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" Apr 21 14:28:24.180586 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:24.180533 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1779e5fc-59fd-4957-a145-e08e128869c7-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-fk6qh\" (UID: \"1779e5fc-59fd-4957-a145-e08e128869c7\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" Apr 21 14:28:24.183119 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:24.183088 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1779e5fc-59fd-4957-a145-e08e128869c7-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-fk6qh\" (UID: \"1779e5fc-59fd-4957-a145-e08e128869c7\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" Apr 21 14:28:24.296268 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:24.296189 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" Apr 21 14:28:24.415457 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:24.415437 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh"] Apr 21 14:28:24.417951 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:24.417910 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1779e5fc_59fd_4957_a145_e08e128869c7.slice/crio-8e0acf1baab9c03a4df7ee51bd65a471d25d07c3ee06848ec1aea77972b4c141 WatchSource:0}: Error finding container 8e0acf1baab9c03a4df7ee51bd65a471d25d07c3ee06848ec1aea77972b4c141: Status 404 returned error can't find the container with id 8e0acf1baab9c03a4df7ee51bd65a471d25d07c3ee06848ec1aea77972b4c141 Apr 21 14:28:24.802114 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:24.802083 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" event={"ID":"1779e5fc-59fd-4957-a145-e08e128869c7","Type":"ContainerStarted","Data":"8e0acf1baab9c03a4df7ee51bd65a471d25d07c3ee06848ec1aea77972b4c141"} Apr 21 14:28:25.806545 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:25.806506 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" event={"ID":"1779e5fc-59fd-4957-a145-e08e128869c7","Type":"ContainerStarted","Data":"1f1a3aa48ea3ec7273ad6c3b1f7d28a89d09019a6b0fa3ad0c2bb8b5d0034dfd"} Apr 21 14:28:25.806989 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:25.806670 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" Apr 21 14:28:25.811509 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:25.811487 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" Apr 21 14:28:25.823615 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:25.823560 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-fk6qh" podStartSLOduration=1.541221451 podStartE2EDuration="2.82354832s" podCreationTimestamp="2026-04-21 14:28:23 +0000 UTC" firstStartedPulling="2026-04-21 14:28:24.419759341 +0000 UTC m=+137.877243827" lastFinishedPulling="2026-04-21 14:28:25.702086197 +0000 UTC m=+139.159570696" observedRunningTime="2026-04-21 14:28:25.822451939 +0000 UTC m=+139.279936446" watchObservedRunningTime="2026-04-21 14:28:25.82354832 +0000 UTC m=+139.281032826" Apr 21 14:28:28.371394 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.371353 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-6f764fb56d-8p9zs"] Apr 21 14:28:28.373750 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.373726 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.376122 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.376098 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 21 14:28:28.376122 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.376108 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 21 14:28:28.377126 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.377104 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 21 14:28:28.377126 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.377134 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 21 14:28:28.377315 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.377142 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 21 14:28:28.377315 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.377176 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 21 14:28:28.377315 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.377137 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-d4rzb\"" Apr 21 14:28:28.377452 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.377402 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 21 14:28:28.386057 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.386032 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f764fb56d-8p9zs"] Apr 21 14:28:28.414246 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.414214 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-config\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.414352 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.414267 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-oauth-config\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.414352 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.414291 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-service-ca\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.414449 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.414345 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-oauth-serving-cert\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.414449 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.414372 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-serving-cert\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.414449 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.414396 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thtsg\" (UniqueName: \"kubernetes.io/projected/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-kube-api-access-thtsg\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.515034 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.514996 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-config\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.515199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.515056 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-oauth-config\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.515199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.515090 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-service-ca\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.515199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.515120 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-oauth-serving-cert\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.515347 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.515272 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-serving-cert\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.515347 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.515332 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-thtsg\" (UniqueName: \"kubernetes.io/projected/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-kube-api-access-thtsg\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.515750 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.515718 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-config\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.516401 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.516377 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-oauth-serving-cert\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.516517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.516498 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-service-ca\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.517780 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.517761 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-oauth-config\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.518453 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.518435 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-serving-cert\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.523712 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.523692 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-thtsg\" (UniqueName: \"kubernetes.io/projected/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-kube-api-access-thtsg\") pod \"console-6f764fb56d-8p9zs\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.685145 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.685104 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:28.806885 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.806859 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f764fb56d-8p9zs"] Apr 21 14:28:28.809152 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:28.809124 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda29dcab9_62c7_48ca_b688_8bc62a43d8ea.slice/crio-be9d6b6d3d04a3b3114c09c2b1b7e0be93270fc4bcb5efdf4c38acd583fc8673 WatchSource:0}: Error finding container be9d6b6d3d04a3b3114c09c2b1b7e0be93270fc4bcb5efdf4c38acd583fc8673: Status 404 returned error can't find the container with id be9d6b6d3d04a3b3114c09c2b1b7e0be93270fc4bcb5efdf4c38acd583fc8673 Apr 21 14:28:28.814363 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:28.814335 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f764fb56d-8p9zs" event={"ID":"a29dcab9-62c7-48ca-b688-8bc62a43d8ea","Type":"ContainerStarted","Data":"be9d6b6d3d04a3b3114c09c2b1b7e0be93270fc4bcb5efdf4c38acd583fc8673"} Apr 21 14:28:32.460748 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:32.460722 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:28:32.827049 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:32.826959 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f764fb56d-8p9zs" event={"ID":"a29dcab9-62c7-48ca-b688-8bc62a43d8ea","Type":"ContainerStarted","Data":"b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b"} Apr 21 14:28:32.845608 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:32.845527 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6f764fb56d-8p9zs" podStartSLOduration=1.533732792 podStartE2EDuration="4.845507891s" podCreationTimestamp="2026-04-21 14:28:28 +0000 UTC" firstStartedPulling="2026-04-21 14:28:28.811042944 +0000 UTC m=+142.268527429" lastFinishedPulling="2026-04-21 14:28:32.12281804 +0000 UTC m=+145.580302528" observedRunningTime="2026-04-21 14:28:32.844741384 +0000 UTC m=+146.302225912" watchObservedRunningTime="2026-04-21 14:28:32.845507891 +0000 UTC m=+146.302992399" Apr 21 14:28:34.762882 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:34.762854 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-79f755579b-8wxj4" Apr 21 14:28:37.476929 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.476888 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" podUID="d85ed974-2064-4d7f-a10c-ff2a82315054" containerName="registry" containerID="cri-o://6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c" gracePeriod=30 Apr 21 14:28:37.719513 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.719487 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:28:37.756491 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.756417 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-5479846bfb-s85np"] Apr 21 14:28:37.756752 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.756736 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d85ed974-2064-4d7f-a10c-ff2a82315054" containerName="registry" Apr 21 14:28:37.756805 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.756756 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85ed974-2064-4d7f-a10c-ff2a82315054" containerName="registry" Apr 21 14:28:37.756841 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.756806 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="d85ed974-2064-4d7f-a10c-ff2a82315054" containerName="registry" Apr 21 14:28:37.758655 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.758637 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.765410 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.765386 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 21 14:28:37.769754 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.769726 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5479846bfb-s85np"] Apr 21 14:28:37.800121 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800092 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d85ed974-2064-4d7f-a10c-ff2a82315054-ca-trust-extracted\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800265 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800129 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-trusted-ca\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800265 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800152 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-bound-sa-token\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800265 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800186 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800265 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800221 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-image-registry-private-configuration\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800442 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800285 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44l2p\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-kube-api-access-44l2p\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800442 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800334 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-installation-pull-secrets\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800442 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800364 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-certificates\") pod \"d85ed974-2064-4d7f-a10c-ff2a82315054\" (UID: \"d85ed974-2064-4d7f-a10c-ff2a82315054\") " Apr 21 14:28:37.800623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800475 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stxf8\" (UniqueName: \"kubernetes.io/projected/60061580-0484-4662-a943-8cfc1e22a4df-kube-api-access-stxf8\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.800623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800508 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-serving-cert\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.800623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800552 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:28:37.800623 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800611 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-trusted-ca-bundle\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.800828 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800669 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-oauth-serving-cert\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.800828 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800761 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-oauth-config\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.800828 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800802 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-console-config\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.800974 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800827 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-service-ca\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.800974 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.800873 2580 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-trusted-ca\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.801131 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.801104 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:28:37.802858 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.802811 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:28:37.802858 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.802840 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:28:37.803031 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.802947 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:28:37.803464 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.803441 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-kube-api-access-44l2p" (OuterVolumeSpecName: "kube-api-access-44l2p") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "kube-api-access-44l2p". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:28:37.803518 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.803451 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:28:37.809093 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.809071 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d85ed974-2064-4d7f-a10c-ff2a82315054-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d85ed974-2064-4d7f-a10c-ff2a82315054" (UID: "d85ed974-2064-4d7f-a10c-ff2a82315054"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 14:28:37.846691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.842495 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" event={"ID":"d85ed974-2064-4d7f-a10c-ff2a82315054","Type":"ContainerDied","Data":"6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c"} Apr 21 14:28:37.846691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.842557 2580 scope.go:117] "RemoveContainer" containerID="6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c" Apr 21 14:28:37.846691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.842800 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" Apr 21 14:28:37.846691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.842464 2580 generic.go:358] "Generic (PLEG): container finished" podID="d85ed974-2064-4d7f-a10c-ff2a82315054" containerID="6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c" exitCode=0 Apr 21 14:28:37.846691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.842945 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-7c56c698b5-pb9q6" event={"ID":"d85ed974-2064-4d7f-a10c-ff2a82315054","Type":"ContainerDied","Data":"ab6aa87fc09aeec322b58e19e6d2180b3ab0f1c6f670496e89e39e0b18648db2"} Apr 21 14:28:37.855306 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.855289 2580 scope.go:117] "RemoveContainer" containerID="6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c" Apr 21 14:28:37.855534 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:28:37.855513 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c\": container with ID starting with 6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c not found: ID does not exist" containerID="6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c" Apr 21 14:28:37.855616 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.855540 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c"} err="failed to get container status \"6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c\": rpc error: code = NotFound desc = could not find container \"6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c\": container with ID starting with 6c51c4fe6200739341970f610006467b385b8e02b07d5e637d3d605fcf02865c not found: ID does not exist" Apr 21 14:28:37.865594 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.865555 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-7c56c698b5-pb9q6"] Apr 21 14:28:37.870868 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.870838 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-7c56c698b5-pb9q6"] Apr 21 14:28:37.901779 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.901757 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-console-config\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.901874 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.901782 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-service-ca\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.901874 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.901804 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-stxf8\" (UniqueName: \"kubernetes.io/projected/60061580-0484-4662-a943-8cfc1e22a4df-kube-api-access-stxf8\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.901947 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.901925 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-serving-cert\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.902010 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.901994 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-trusted-ca-bundle\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.902058 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902050 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-oauth-serving-cert\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.902149 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902133 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-oauth-config\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.902222 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902191 2580 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-installation-pull-secrets\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.902222 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902208 2580 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-certificates\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.902317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902223 2580 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d85ed974-2064-4d7f-a10c-ff2a82315054-ca-trust-extracted\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.902317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902239 2580 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-bound-sa-token\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.902317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902254 2580 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-registry-tls\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.902317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902270 2580 reconciler_common.go:299] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d85ed974-2064-4d7f-a10c-ff2a82315054-image-registry-private-configuration\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.902317 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902285 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-44l2p\" (UniqueName: \"kubernetes.io/projected/d85ed974-2064-4d7f-a10c-ff2a82315054-kube-api-access-44l2p\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:28:37.902547 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902461 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-service-ca\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.902547 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902524 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-console-config\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.902794 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902770 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-trusted-ca-bundle\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.902984 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.902879 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-oauth-serving-cert\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.904528 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.904513 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-serving-cert\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.904593 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.904529 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-oauth-config\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:37.910387 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:37.910370 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-stxf8\" (UniqueName: \"kubernetes.io/projected/60061580-0484-4662-a943-8cfc1e22a4df-kube-api-access-stxf8\") pod \"console-5479846bfb-s85np\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:38.067630 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.067515 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:38.219364 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.219329 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5479846bfb-s85np"] Apr 21 14:28:38.223761 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:38.223735 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60061580_0484_4662_a943_8cfc1e22a4df.slice/crio-a917c53a49c174177f8d4d281df3bf8b6f9c6e90886b3439a4c3d8313725e418 WatchSource:0}: Error finding container a917c53a49c174177f8d4d281df3bf8b6f9c6e90886b3439a4c3d8313725e418: Status 404 returned error can't find the container with id a917c53a49c174177f8d4d281df3bf8b6f9c6e90886b3439a4c3d8313725e418 Apr 21 14:28:38.685231 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.685195 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:38.685692 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.685289 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:38.689858 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.689838 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:38.848363 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.848322 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5479846bfb-s85np" event={"ID":"60061580-0484-4662-a943-8cfc1e22a4df","Type":"ContainerStarted","Data":"a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954"} Apr 21 14:28:38.848363 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.848366 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5479846bfb-s85np" event={"ID":"60061580-0484-4662-a943-8cfc1e22a4df","Type":"ContainerStarted","Data":"a917c53a49c174177f8d4d281df3bf8b6f9c6e90886b3439a4c3d8313725e418"} Apr 21 14:28:38.852138 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.852108 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:28:38.868677 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:38.868634 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5479846bfb-s85np" podStartSLOduration=1.868590398 podStartE2EDuration="1.868590398s" podCreationTimestamp="2026-04-21 14:28:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:28:38.867695543 +0000 UTC m=+152.325180062" watchObservedRunningTime="2026-04-21 14:28:38.868590398 +0000 UTC m=+152.326074904" Apr 21 14:28:39.245515 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:39.245469 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d85ed974-2064-4d7f-a10c-ff2a82315054" path="/var/lib/kubelet/pods/d85ed974-2064-4d7f-a10c-ff2a82315054/volumes" Apr 21 14:28:42.973979 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:28:42.973937 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-dx8mq" podUID="afea4c93-a952-4362-8775-179f3919eecc" Apr 21 14:28:42.996114 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:28:42.996077 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-tfksq" podUID="5badb5d3-1012-4fbf-81cc-b5190659075e" Apr 21 14:28:43.860485 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:43.860453 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dx8mq" Apr 21 14:28:47.897840 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:47.897804 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:28:47.898219 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:47.897852 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:28:47.900370 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:47.900336 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afea4c93-a952-4362-8775-179f3919eecc-metrics-tls\") pod \"dns-default-dx8mq\" (UID: \"afea4c93-a952-4362-8775-179f3919eecc\") " pod="openshift-dns/dns-default-dx8mq" Apr 21 14:28:47.900370 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:47.900365 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5badb5d3-1012-4fbf-81cc-b5190659075e-cert\") pod \"ingress-canary-tfksq\" (UID: \"5badb5d3-1012-4fbf-81cc-b5190659075e\") " pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:28:48.064948 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.064917 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-88srx\"" Apr 21 14:28:48.068035 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.068016 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:48.068098 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.068052 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:48.072529 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.072503 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dx8mq" Apr 21 14:28:48.072885 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.072861 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:48.199715 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.199656 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dx8mq"] Apr 21 14:28:48.203453 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:48.203425 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafea4c93_a952_4362_8775_179f3919eecc.slice/crio-13501c152751d505423c8aa683f2e8d0e80d41b019838d645bfe3ecaf238a262 WatchSource:0}: Error finding container 13501c152751d505423c8aa683f2e8d0e80d41b019838d645bfe3ecaf238a262: Status 404 returned error can't find the container with id 13501c152751d505423c8aa683f2e8d0e80d41b019838d645bfe3ecaf238a262 Apr 21 14:28:48.875377 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.875341 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dx8mq" event={"ID":"afea4c93-a952-4362-8775-179f3919eecc","Type":"ContainerStarted","Data":"13501c152751d505423c8aa683f2e8d0e80d41b019838d645bfe3ecaf238a262"} Apr 21 14:28:48.880555 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.880528 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:28:48.926654 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:48.926624 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6f764fb56d-8p9zs"] Apr 21 14:28:49.879664 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:49.879548 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dx8mq" event={"ID":"afea4c93-a952-4362-8775-179f3919eecc","Type":"ContainerStarted","Data":"d1f3535f2090a629b82443610a61b69d8571cf2e12d8a55798e08f4005559657"} Apr 21 14:28:49.879664 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:49.879611 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dx8mq" event={"ID":"afea4c93-a952-4362-8775-179f3919eecc","Type":"ContainerStarted","Data":"189306183fdacef0c95c56ff420220590a4c05e274ff0587a8b3a996b975b0cf"} Apr 21 14:28:49.895853 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:49.895668 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dx8mq" podStartSLOduration=129.572455213 podStartE2EDuration="2m10.895650309s" podCreationTimestamp="2026-04-21 14:26:39 +0000 UTC" firstStartedPulling="2026-04-21 14:28:48.205399371 +0000 UTC m=+161.662883856" lastFinishedPulling="2026-04-21 14:28:49.52859445 +0000 UTC m=+162.986078952" observedRunningTime="2026-04-21 14:28:49.895013926 +0000 UTC m=+163.352498432" watchObservedRunningTime="2026-04-21 14:28:49.895650309 +0000 UTC m=+163.353134817" Apr 21 14:28:50.882429 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:50.882398 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-dx8mq" Apr 21 14:28:51.885818 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:51.885782 2580 generic.go:358] "Generic (PLEG): container finished" podID="07d5eb46-986b-4f0e-9f2b-6db3b006c975" containerID="3be06cc8dd517999aafae4a0afc50c45546d1de11783e3b32a303ce2b52a5248" exitCode=0 Apr 21 14:28:51.886300 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:51.885854 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" event={"ID":"07d5eb46-986b-4f0e-9f2b-6db3b006c975","Type":"ContainerDied","Data":"3be06cc8dd517999aafae4a0afc50c45546d1de11783e3b32a303ce2b52a5248"} Apr 21 14:28:51.886300 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:51.886225 2580 scope.go:117] "RemoveContainer" containerID="3be06cc8dd517999aafae4a0afc50c45546d1de11783e3b32a303ce2b52a5248" Apr 21 14:28:52.889949 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:52.889910 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-2lrpd" event={"ID":"07d5eb46-986b-4f0e-9f2b-6db3b006c975","Type":"ContainerStarted","Data":"a2c82dede8522989da9dc55addddcd7e0a6f72a9f1ebb41b024375ec69070225"} Apr 21 14:28:54.240952 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:54.240901 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:28:54.243484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:54.243466 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-pqf56\"" Apr 21 14:28:54.252222 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:54.252204 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tfksq" Apr 21 14:28:54.370516 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:54.370468 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tfksq"] Apr 21 14:28:54.372513 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:28:54.372485 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5badb5d3_1012_4fbf_81cc_b5190659075e.slice/crio-98da0222bb8a63d840159b50a103ce8f8f2e8303a82ddcb155d3ee579cf61a17 WatchSource:0}: Error finding container 98da0222bb8a63d840159b50a103ce8f8f2e8303a82ddcb155d3ee579cf61a17: Status 404 returned error can't find the container with id 98da0222bb8a63d840159b50a103ce8f8f2e8303a82ddcb155d3ee579cf61a17 Apr 21 14:28:54.896871 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:54.896833 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tfksq" event={"ID":"5badb5d3-1012-4fbf-81cc-b5190659075e","Type":"ContainerStarted","Data":"98da0222bb8a63d840159b50a103ce8f8f2e8303a82ddcb155d3ee579cf61a17"} Apr 21 14:28:56.904473 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:56.904434 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tfksq" event={"ID":"5badb5d3-1012-4fbf-81cc-b5190659075e","Type":"ContainerStarted","Data":"5e75f50c38d7a116ec75e9365788217090b2f3e864a0058dd90e0c38bece6424"} Apr 21 14:28:56.920740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:28:56.920686 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-tfksq" podStartSLOduration=136.030927671 podStartE2EDuration="2m17.920669547s" podCreationTimestamp="2026-04-21 14:26:39 +0000 UTC" firstStartedPulling="2026-04-21 14:28:54.376918467 +0000 UTC m=+167.834402952" lastFinishedPulling="2026-04-21 14:28:56.266660341 +0000 UTC m=+169.724144828" observedRunningTime="2026-04-21 14:28:56.920059763 +0000 UTC m=+170.377544270" watchObservedRunningTime="2026-04-21 14:28:56.920669547 +0000 UTC m=+170.378154054" Apr 21 14:29:00.888650 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:00.888618 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dx8mq" Apr 21 14:29:08.940038 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:08.940005 2580 generic.go:358] "Generic (PLEG): container finished" podID="4c20781a-bb09-4f47-8dcd-d4d713caac69" containerID="abccd478d277afe4e0461e00ec21f6821ac0e3c69942deae55c86efcc9036296" exitCode=0 Apr 21 14:29:08.940429 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:08.940083 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-ckmd7" event={"ID":"4c20781a-bb09-4f47-8dcd-d4d713caac69","Type":"ContainerDied","Data":"abccd478d277afe4e0461e00ec21f6821ac0e3c69942deae55c86efcc9036296"} Apr 21 14:29:08.940429 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:08.940411 2580 scope.go:117] "RemoveContainer" containerID="abccd478d277afe4e0461e00ec21f6821ac0e3c69942deae55c86efcc9036296" Apr 21 14:29:09.944777 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:09.944743 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-ckmd7" event={"ID":"4c20781a-bb09-4f47-8dcd-d4d713caac69","Type":"ContainerStarted","Data":"f9c6ea50c8f7ebc9c3745bfe81a0747244d1608f9be7343cf6e3182a24e76155"} Apr 21 14:29:13.950602 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:13.950534 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-6f764fb56d-8p9zs" podUID="a29dcab9-62c7-48ca-b688-8bc62a43d8ea" containerName="console" containerID="cri-o://b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b" gracePeriod=15 Apr 21 14:29:14.188811 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.188786 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6f764fb56d-8p9zs_a29dcab9-62c7-48ca-b688-8bc62a43d8ea/console/0.log" Apr 21 14:29:14.188940 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.188851 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:29:14.216019 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.215937 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thtsg\" (UniqueName: \"kubernetes.io/projected/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-kube-api-access-thtsg\") pod \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " Apr 21 14:29:14.216019 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.215980 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-oauth-config\") pod \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " Apr 21 14:29:14.216225 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.216025 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-service-ca\") pod \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " Apr 21 14:29:14.216225 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.216074 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-oauth-serving-cert\") pod \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " Apr 21 14:29:14.216225 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.216101 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-config\") pod \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " Apr 21 14:29:14.216225 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.216126 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-serving-cert\") pod \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\" (UID: \"a29dcab9-62c7-48ca-b688-8bc62a43d8ea\") " Apr 21 14:29:14.216682 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.216635 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-service-ca" (OuterVolumeSpecName: "service-ca") pod "a29dcab9-62c7-48ca-b688-8bc62a43d8ea" (UID: "a29dcab9-62c7-48ca-b688-8bc62a43d8ea"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:29:14.216682 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.216662 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a29dcab9-62c7-48ca-b688-8bc62a43d8ea" (UID: "a29dcab9-62c7-48ca-b688-8bc62a43d8ea"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:29:14.216859 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.216747 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-config" (OuterVolumeSpecName: "console-config") pod "a29dcab9-62c7-48ca-b688-8bc62a43d8ea" (UID: "a29dcab9-62c7-48ca-b688-8bc62a43d8ea"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:29:14.218490 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.218468 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a29dcab9-62c7-48ca-b688-8bc62a43d8ea" (UID: "a29dcab9-62c7-48ca-b688-8bc62a43d8ea"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:29:14.218949 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.218928 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-kube-api-access-thtsg" (OuterVolumeSpecName: "kube-api-access-thtsg") pod "a29dcab9-62c7-48ca-b688-8bc62a43d8ea" (UID: "a29dcab9-62c7-48ca-b688-8bc62a43d8ea"). InnerVolumeSpecName "kube-api-access-thtsg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:29:14.219003 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.218930 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a29dcab9-62c7-48ca-b688-8bc62a43d8ea" (UID: "a29dcab9-62c7-48ca-b688-8bc62a43d8ea"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:29:14.317069 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.317018 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-thtsg\" (UniqueName: \"kubernetes.io/projected/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-kube-api-access-thtsg\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:29:14.317069 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.317065 2580 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-oauth-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:29:14.317069 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.317076 2580 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-service-ca\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:29:14.317069 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.317085 2580 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-oauth-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:29:14.317347 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.317096 2580 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:29:14.317347 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.317107 2580 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a29dcab9-62c7-48ca-b688-8bc62a43d8ea-console-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:29:14.963435 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.963407 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6f764fb56d-8p9zs_a29dcab9-62c7-48ca-b688-8bc62a43d8ea/console/0.log" Apr 21 14:29:14.963886 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.963451 2580 generic.go:358] "Generic (PLEG): container finished" podID="a29dcab9-62c7-48ca-b688-8bc62a43d8ea" containerID="b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b" exitCode=2 Apr 21 14:29:14.963886 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.963484 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f764fb56d-8p9zs" event={"ID":"a29dcab9-62c7-48ca-b688-8bc62a43d8ea","Type":"ContainerDied","Data":"b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b"} Apr 21 14:29:14.963886 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.963506 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f764fb56d-8p9zs" event={"ID":"a29dcab9-62c7-48ca-b688-8bc62a43d8ea","Type":"ContainerDied","Data":"be9d6b6d3d04a3b3114c09c2b1b7e0be93270fc4bcb5efdf4c38acd583fc8673"} Apr 21 14:29:14.963886 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.963520 2580 scope.go:117] "RemoveContainer" containerID="b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b" Apr 21 14:29:14.963886 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.963523 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f764fb56d-8p9zs" Apr 21 14:29:14.972330 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.972311 2580 scope.go:117] "RemoveContainer" containerID="b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b" Apr 21 14:29:14.972626 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:29:14.972605 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b\": container with ID starting with b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b not found: ID does not exist" containerID="b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b" Apr 21 14:29:14.972701 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.972639 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b"} err="failed to get container status \"b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b\": rpc error: code = NotFound desc = could not find container \"b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b\": container with ID starting with b81878b4e8b9bdd7e1ca413bb07c88234bc18f06e0cdfad95f295d1577cd501b not found: ID does not exist" Apr 21 14:29:14.984297 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.984268 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6f764fb56d-8p9zs"] Apr 21 14:29:14.990773 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:14.990747 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-6f764fb56d-8p9zs"] Apr 21 14:29:15.245358 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:15.245264 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29dcab9-62c7-48ca-b688-8bc62a43d8ea" path="/var/lib/kubelet/pods/a29dcab9-62c7-48ca-b688-8bc62a43d8ea/volumes" Apr 21 14:29:39.038733 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:39.038702 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:29:39.039171 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:39.038743 2580 generic.go:358] "Generic (PLEG): container finished" podID="f355a71a-6279-4fa0-9123-6a056ee88f72" containerID="ffd6c44893184d4d13c4696105e1e8aefa35ea88ed72d87494e6eadb4e2f6ad0" exitCode=2 Apr 21 14:29:39.039171 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:39.038778 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" event={"ID":"f355a71a-6279-4fa0-9123-6a056ee88f72","Type":"ContainerDied","Data":"ffd6c44893184d4d13c4696105e1e8aefa35ea88ed72d87494e6eadb4e2f6ad0"} Apr 21 14:29:39.039171 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:39.039085 2580 scope.go:117] "RemoveContainer" containerID="ffd6c44893184d4d13c4696105e1e8aefa35ea88ed72d87494e6eadb4e2f6ad0" Apr 21 14:29:40.043988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:40.043960 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:29:40.044360 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:40.044010 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-pn2df" event={"ID":"f355a71a-6279-4fa0-9123-6a056ee88f72","Type":"ContainerStarted","Data":"117c2042558e99e32093f0fab7232d1ec70aa940614b21c163bc64cd20d0f127"} Apr 21 14:29:52.574303 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.574270 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-8c9c4657f-jg27h"] Apr 21 14:29:52.574826 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.574714 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a29dcab9-62c7-48ca-b688-8bc62a43d8ea" containerName="console" Apr 21 14:29:52.574826 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.574734 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29dcab9-62c7-48ca-b688-8bc62a43d8ea" containerName="console" Apr 21 14:29:52.574826 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.574817 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="a29dcab9-62c7-48ca-b688-8bc62a43d8ea" containerName="console" Apr 21 14:29:52.577738 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.577717 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.590894 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.590867 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8c9c4657f-jg27h"] Apr 21 14:29:52.744873 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.744838 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75ms2\" (UniqueName: \"kubernetes.io/projected/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-kube-api-access-75ms2\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.744873 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.744878 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-trusted-ca-bundle\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.745084 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.744909 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-oauth-config\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.745084 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.745015 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-oauth-serving-cert\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.745084 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.745048 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-serving-cert\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.745182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.745098 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-config\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.745182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.745125 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-service-ca\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.845609 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.845488 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-serving-cert\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.845609 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.845531 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-config\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.845609 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.845556 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-service-ca\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.845904 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.845700 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-75ms2\" (UniqueName: \"kubernetes.io/projected/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-kube-api-access-75ms2\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.845904 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.845759 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-trusted-ca-bundle\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.845904 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.845807 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-oauth-config\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.845904 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.845858 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-oauth-serving-cert\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.846384 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.846333 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-service-ca\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.846384 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.846373 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-config\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.846596 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.846529 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-oauth-serving-cert\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.846933 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.846910 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-trusted-ca-bundle\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.848296 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.848272 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-serving-cert\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.848383 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.848276 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-oauth-config\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.857603 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.857582 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-75ms2\" (UniqueName: \"kubernetes.io/projected/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-kube-api-access-75ms2\") pod \"console-8c9c4657f-jg27h\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:52.888517 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:52.888476 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:29:53.010740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:53.010715 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8c9c4657f-jg27h"] Apr 21 14:29:53.013204 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:29:53.013177 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a3b5fe8_b6c0_4431_bddc_e4f190739eb6.slice/crio-0864b9a4dd806b3873832b7c76d49fabb0e48a6a36b0a41eb60686c69adc5603 WatchSource:0}: Error finding container 0864b9a4dd806b3873832b7c76d49fabb0e48a6a36b0a41eb60686c69adc5603: Status 404 returned error can't find the container with id 0864b9a4dd806b3873832b7c76d49fabb0e48a6a36b0a41eb60686c69adc5603 Apr 21 14:29:53.083756 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:53.083724 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8c9c4657f-jg27h" event={"ID":"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6","Type":"ContainerStarted","Data":"b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104"} Apr 21 14:29:53.083871 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:53.083762 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8c9c4657f-jg27h" event={"ID":"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6","Type":"ContainerStarted","Data":"0864b9a4dd806b3873832b7c76d49fabb0e48a6a36b0a41eb60686c69adc5603"} Apr 21 14:29:53.102997 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:29:53.102890 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-8c9c4657f-jg27h" podStartSLOduration=1.102871568 podStartE2EDuration="1.102871568s" podCreationTimestamp="2026-04-21 14:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:29:53.101130123 +0000 UTC m=+226.558614630" watchObservedRunningTime="2026-04-21 14:29:53.102871568 +0000 UTC m=+226.560356076" Apr 21 14:30:02.889118 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:02.889074 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:30:02.889530 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:02.889148 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:30:02.893992 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:02.893962 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:30:03.114919 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:03.114891 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:30:03.163095 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:03.163000 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5479846bfb-s85np"] Apr 21 14:30:28.187861 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.187803 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-5479846bfb-s85np" podUID="60061580-0484-4662-a943-8cfc1e22a4df" containerName="console" containerID="cri-o://a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954" gracePeriod=15 Apr 21 14:30:28.428781 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.428760 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5479846bfb-s85np_60061580-0484-4662-a943-8cfc1e22a4df/console/0.log" Apr 21 14:30:28.428893 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.428828 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:30:28.527918 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.527828 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stxf8\" (UniqueName: \"kubernetes.io/projected/60061580-0484-4662-a943-8cfc1e22a4df-kube-api-access-stxf8\") pod \"60061580-0484-4662-a943-8cfc1e22a4df\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " Apr 21 14:30:28.527918 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.527898 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-service-ca\") pod \"60061580-0484-4662-a943-8cfc1e22a4df\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " Apr 21 14:30:28.528141 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.527994 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-oauth-config\") pod \"60061580-0484-4662-a943-8cfc1e22a4df\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " Apr 21 14:30:28.528141 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528039 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-oauth-serving-cert\") pod \"60061580-0484-4662-a943-8cfc1e22a4df\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " Apr 21 14:30:28.528141 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528059 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-trusted-ca-bundle\") pod \"60061580-0484-4662-a943-8cfc1e22a4df\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " Apr 21 14:30:28.528141 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528114 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-serving-cert\") pod \"60061580-0484-4662-a943-8cfc1e22a4df\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " Apr 21 14:30:28.528141 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528137 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-console-config\") pod \"60061580-0484-4662-a943-8cfc1e22a4df\" (UID: \"60061580-0484-4662-a943-8cfc1e22a4df\") " Apr 21 14:30:28.528404 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528370 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-service-ca" (OuterVolumeSpecName: "service-ca") pod "60061580-0484-4662-a943-8cfc1e22a4df" (UID: "60061580-0484-4662-a943-8cfc1e22a4df"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:30:28.528460 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528391 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "60061580-0484-4662-a943-8cfc1e22a4df" (UID: "60061580-0484-4662-a943-8cfc1e22a4df"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:30:28.528545 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528517 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "60061580-0484-4662-a943-8cfc1e22a4df" (UID: "60061580-0484-4662-a943-8cfc1e22a4df"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:30:28.528689 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.528667 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-console-config" (OuterVolumeSpecName: "console-config") pod "60061580-0484-4662-a943-8cfc1e22a4df" (UID: "60061580-0484-4662-a943-8cfc1e22a4df"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:30:28.530167 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.530143 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "60061580-0484-4662-a943-8cfc1e22a4df" (UID: "60061580-0484-4662-a943-8cfc1e22a4df"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:30:28.530313 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.530285 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "60061580-0484-4662-a943-8cfc1e22a4df" (UID: "60061580-0484-4662-a943-8cfc1e22a4df"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:30:28.530371 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.530296 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60061580-0484-4662-a943-8cfc1e22a4df-kube-api-access-stxf8" (OuterVolumeSpecName: "kube-api-access-stxf8") pod "60061580-0484-4662-a943-8cfc1e22a4df" (UID: "60061580-0484-4662-a943-8cfc1e22a4df"). InnerVolumeSpecName "kube-api-access-stxf8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:30:28.629264 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.629225 2580 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-service-ca\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:30:28.629264 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.629257 2580 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-oauth-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:30:28.629264 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.629269 2580 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-oauth-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:30:28.629492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.629279 2580 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-trusted-ca-bundle\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:30:28.629492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.629288 2580 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60061580-0484-4662-a943-8cfc1e22a4df-console-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:30:28.629492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.629296 2580 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60061580-0484-4662-a943-8cfc1e22a4df-console-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:30:28.629492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:28.629305 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-stxf8\" (UniqueName: \"kubernetes.io/projected/60061580-0484-4662-a943-8cfc1e22a4df-kube-api-access-stxf8\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:30:29.186492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.186466 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5479846bfb-s85np_60061580-0484-4662-a943-8cfc1e22a4df/console/0.log" Apr 21 14:30:29.186722 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.186506 2580 generic.go:358] "Generic (PLEG): container finished" podID="60061580-0484-4662-a943-8cfc1e22a4df" containerID="a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954" exitCode=2 Apr 21 14:30:29.186722 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.186603 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5479846bfb-s85np" Apr 21 14:30:29.186722 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.186603 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5479846bfb-s85np" event={"ID":"60061580-0484-4662-a943-8cfc1e22a4df","Type":"ContainerDied","Data":"a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954"} Apr 21 14:30:29.186722 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.186642 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5479846bfb-s85np" event={"ID":"60061580-0484-4662-a943-8cfc1e22a4df","Type":"ContainerDied","Data":"a917c53a49c174177f8d4d281df3bf8b6f9c6e90886b3439a4c3d8313725e418"} Apr 21 14:30:29.186722 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.186657 2580 scope.go:117] "RemoveContainer" containerID="a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954" Apr 21 14:30:29.195177 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.195079 2580 scope.go:117] "RemoveContainer" containerID="a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954" Apr 21 14:30:29.195398 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:30:29.195304 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954\": container with ID starting with a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954 not found: ID does not exist" containerID="a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954" Apr 21 14:30:29.195398 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.195329 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954"} err="failed to get container status \"a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954\": rpc error: code = NotFound desc = could not find container \"a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954\": container with ID starting with a89cadf99d49102d8f1403d9a5462b196a8ee8b6cfb1abbbd4ce227085e51954 not found: ID does not exist" Apr 21 14:30:29.207250 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.207223 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5479846bfb-s85np"] Apr 21 14:30:29.210952 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.210924 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5479846bfb-s85np"] Apr 21 14:30:29.246745 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:30:29.246709 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60061580-0484-4662-a943-8cfc1e22a4df" path="/var/lib/kubelet/pods/60061580-0484-4662-a943-8cfc1e22a4df/volumes" Apr 21 14:31:07.109774 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:07.109744 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:31:07.110314 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:07.109754 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:31:07.111452 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:07.111431 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:31:07.111602 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:07.111436 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:31:07.114583 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:07.114542 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:31:07.114746 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:07.114544 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:31:07.121820 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:07.121802 2580 kubelet.go:1628] "Image garbage collection succeeded" Apr 21 14:31:23.142147 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.142070 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-ztmwm"] Apr 21 14:31:23.144478 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.142384 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="60061580-0484-4662-a943-8cfc1e22a4df" containerName="console" Apr 21 14:31:23.144478 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.142398 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="60061580-0484-4662-a943-8cfc1e22a4df" containerName="console" Apr 21 14:31:23.144478 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.142456 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="60061580-0484-4662-a943-8cfc1e22a4df" containerName="console" Apr 21 14:31:23.145461 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.145437 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.147643 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.147622 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 21 14:31:23.153251 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.153226 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-ztmwm"] Apr 21 14:31:23.253694 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.253658 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-dd4f9cd68-nknr5"] Apr 21 14:31:23.256797 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.256771 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.266201 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.266178 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dd4f9cd68-nknr5"] Apr 21 14:31:23.268430 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.268410 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/ac750e76-e718-4228-a16a-ea303bb3ed02-dbus\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.268530 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.268440 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/ac750e76-e718-4228-a16a-ea303bb3ed02-kubelet-config\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.268530 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.268459 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/ac750e76-e718-4228-a16a-ea303bb3ed02-original-pull-secret\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.369307 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369277 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/ac750e76-e718-4228-a16a-ea303bb3ed02-kubelet-config\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.369307 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369311 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/ac750e76-e718-4228-a16a-ea303bb3ed02-original-pull-secret\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369340 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjslb\" (UniqueName: \"kubernetes.io/projected/12116f2b-7c82-4ce8-981d-d10e7fda1763-kube-api-access-wjslb\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369380 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-oauth-config\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369395 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-config\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369413 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-trusted-ca-bundle\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369411 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/ac750e76-e718-4228-a16a-ea303bb3ed02-kubelet-config\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369483 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-serving-cert\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369502 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-oauth-serving-cert\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369533 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-service-ca\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.369599 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369558 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/ac750e76-e718-4228-a16a-ea303bb3ed02-dbus\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.369962 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.369738 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/ac750e76-e718-4228-a16a-ea303bb3ed02-dbus\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.371842 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.371821 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/ac750e76-e718-4228-a16a-ea303bb3ed02-original-pull-secret\") pod \"global-pull-secret-syncer-ztmwm\" (UID: \"ac750e76-e718-4228-a16a-ea303bb3ed02\") " pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.455304 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.455264 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-ztmwm" Apr 21 14:31:23.470349 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.470318 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-trusted-ca-bundle\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.470491 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.470373 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-serving-cert\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.470553 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.470504 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-oauth-serving-cert\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.470643 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.470557 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-service-ca\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.470696 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.470656 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wjslb\" (UniqueName: \"kubernetes.io/projected/12116f2b-7c82-4ce8-981d-d10e7fda1763-kube-api-access-wjslb\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.470747 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.470730 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-oauth-config\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.470799 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.470753 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-config\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.471284 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.471235 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-oauth-serving-cert\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.471395 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.471304 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-trusted-ca-bundle\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.471608 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.471582 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-service-ca\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.472198 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.472172 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-config\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.474725 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.474705 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-serving-cert\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.474836 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.474751 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-oauth-config\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.480705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.480658 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjslb\" (UniqueName: \"kubernetes.io/projected/12116f2b-7c82-4ce8-981d-d10e7fda1763-kube-api-access-wjslb\") pod \"console-dd4f9cd68-nknr5\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.566013 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.565976 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:23.580986 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.580945 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-ztmwm"] Apr 21 14:31:23.585356 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:31:23.585329 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac750e76_e718_4228_a16a_ea303bb3ed02.slice/crio-2db302a46a7126400b3a140198392ecdd1fc27fb3848585434fa44e5de2e8036 WatchSource:0}: Error finding container 2db302a46a7126400b3a140198392ecdd1fc27fb3848585434fa44e5de2e8036: Status 404 returned error can't find the container with id 2db302a46a7126400b3a140198392ecdd1fc27fb3848585434fa44e5de2e8036 Apr 21 14:31:23.587429 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.587411 2580 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 14:31:23.694934 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:23.694911 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dd4f9cd68-nknr5"] Apr 21 14:31:23.696776 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:31:23.696749 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12116f2b_7c82_4ce8_981d_d10e7fda1763.slice/crio-96ff4bb24bf449b0753dfa25cdebbd2ee9a5bcb3b99de201f21823f83b874188 WatchSource:0}: Error finding container 96ff4bb24bf449b0753dfa25cdebbd2ee9a5bcb3b99de201f21823f83b874188: Status 404 returned error can't find the container with id 96ff4bb24bf449b0753dfa25cdebbd2ee9a5bcb3b99de201f21823f83b874188 Apr 21 14:31:24.336329 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:24.336290 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dd4f9cd68-nknr5" event={"ID":"12116f2b-7c82-4ce8-981d-d10e7fda1763","Type":"ContainerStarted","Data":"ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222"} Apr 21 14:31:24.336329 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:24.336326 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dd4f9cd68-nknr5" event={"ID":"12116f2b-7c82-4ce8-981d-d10e7fda1763","Type":"ContainerStarted","Data":"96ff4bb24bf449b0753dfa25cdebbd2ee9a5bcb3b99de201f21823f83b874188"} Apr 21 14:31:24.337281 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:24.337251 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-ztmwm" event={"ID":"ac750e76-e718-4228-a16a-ea303bb3ed02","Type":"ContainerStarted","Data":"2db302a46a7126400b3a140198392ecdd1fc27fb3848585434fa44e5de2e8036"} Apr 21 14:31:24.353709 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:24.353660 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-dd4f9cd68-nknr5" podStartSLOduration=1.353645591 podStartE2EDuration="1.353645591s" podCreationTimestamp="2026-04-21 14:31:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:31:24.352093799 +0000 UTC m=+317.809578321" watchObservedRunningTime="2026-04-21 14:31:24.353645591 +0000 UTC m=+317.811130098" Apr 21 14:31:28.350823 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:28.350788 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-ztmwm" event={"ID":"ac750e76-e718-4228-a16a-ea303bb3ed02","Type":"ContainerStarted","Data":"b489fc3c5e9fae1ed6743a94ad1d0a1c0a3b128893e4bdb8bc4fc392ab9d362d"} Apr 21 14:31:28.366266 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:28.366209 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-ztmwm" podStartSLOduration=0.754701473 podStartE2EDuration="5.366195493s" podCreationTimestamp="2026-04-21 14:31:23 +0000 UTC" firstStartedPulling="2026-04-21 14:31:23.587548681 +0000 UTC m=+317.045033166" lastFinishedPulling="2026-04-21 14:31:28.199042685 +0000 UTC m=+321.656527186" observedRunningTime="2026-04-21 14:31:28.364292072 +0000 UTC m=+321.821776583" watchObservedRunningTime="2026-04-21 14:31:28.366195493 +0000 UTC m=+321.823679999" Apr 21 14:31:33.566664 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:33.566610 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:33.566664 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:33.566676 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:33.571486 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:33.571462 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:34.372627 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:34.372598 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:31:34.416810 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:34.416775 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8c9c4657f-jg27h"] Apr 21 14:31:59.442009 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.441901 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-8c9c4657f-jg27h" podUID="3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" containerName="console" containerID="cri-o://b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104" gracePeriod=15 Apr 21 14:31:59.681319 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.681299 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8c9c4657f-jg27h_3a3b5fe8-b6c0-4431-bddc-e4f190739eb6/console/0.log" Apr 21 14:31:59.681432 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.681362 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:31:59.785469 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785363 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-serving-cert\") pod \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " Apr 21 14:31:59.785469 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785429 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-trusted-ca-bundle\") pod \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " Apr 21 14:31:59.785469 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785465 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-service-ca\") pod \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " Apr 21 14:31:59.785794 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785492 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-config\") pod \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " Apr 21 14:31:59.785794 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785518 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75ms2\" (UniqueName: \"kubernetes.io/projected/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-kube-api-access-75ms2\") pod \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " Apr 21 14:31:59.785794 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785540 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-oauth-serving-cert\") pod \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " Apr 21 14:31:59.785794 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785666 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-oauth-config\") pod \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\" (UID: \"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6\") " Apr 21 14:31:59.786013 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785932 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" (UID: "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:31:59.786013 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.785955 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" (UID: "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:31:59.786103 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.786018 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-service-ca" (OuterVolumeSpecName: "service-ca") pod "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" (UID: "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:31:59.786168 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.786118 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-config" (OuterVolumeSpecName: "console-config") pod "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" (UID: "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:31:59.787945 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.787916 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" (UID: "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:31:59.788038 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.788009 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-kube-api-access-75ms2" (OuterVolumeSpecName: "kube-api-access-75ms2") pod "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" (UID: "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6"). InnerVolumeSpecName "kube-api-access-75ms2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:31:59.788038 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.788019 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" (UID: "3a3b5fe8-b6c0-4431-bddc-e4f190739eb6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:31:59.886272 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.886232 2580 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-oauth-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:31:59.886272 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.886268 2580 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-oauth-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:31:59.886272 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.886281 2580 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:31:59.886558 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.886293 2580 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-trusted-ca-bundle\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:31:59.886558 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.886307 2580 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-service-ca\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:31:59.886558 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.886319 2580 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-console-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:31:59.886558 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:31:59.886330 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-75ms2\" (UniqueName: \"kubernetes.io/projected/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6-kube-api-access-75ms2\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:32:00.437524 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.437497 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8c9c4657f-jg27h_3a3b5fe8-b6c0-4431-bddc-e4f190739eb6/console/0.log" Apr 21 14:32:00.437718 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.437537 2580 generic.go:358] "Generic (PLEG): container finished" podID="3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" containerID="b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104" exitCode=2 Apr 21 14:32:00.437718 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.437594 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8c9c4657f-jg27h" event={"ID":"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6","Type":"ContainerDied","Data":"b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104"} Apr 21 14:32:00.437718 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.437617 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8c9c4657f-jg27h" event={"ID":"3a3b5fe8-b6c0-4431-bddc-e4f190739eb6","Type":"ContainerDied","Data":"0864b9a4dd806b3873832b7c76d49fabb0e48a6a36b0a41eb60686c69adc5603"} Apr 21 14:32:00.437718 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.437631 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8c9c4657f-jg27h" Apr 21 14:32:00.437718 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.437641 2580 scope.go:117] "RemoveContainer" containerID="b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104" Apr 21 14:32:00.448863 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.448667 2580 scope.go:117] "RemoveContainer" containerID="b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104" Apr 21 14:32:00.449116 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:32:00.448993 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104\": container with ID starting with b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104 not found: ID does not exist" containerID="b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104" Apr 21 14:32:00.449116 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.449034 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104"} err="failed to get container status \"b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104\": rpc error: code = NotFound desc = could not find container \"b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104\": container with ID starting with b3037b7fd207a623d93a1301842c5cb56ced78985b4e96d1cd1840d49620d104 not found: ID does not exist" Apr 21 14:32:00.457644 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.457615 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8c9c4657f-jg27h"] Apr 21 14:32:00.461786 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:00.461766 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-8c9c4657f-jg27h"] Apr 21 14:32:01.245376 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:32:01.245329 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" path="/var/lib/kubelet/pods/3a3b5fe8-b6c0-4431-bddc-e4f190739eb6/volumes" Apr 21 14:34:55.078665 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.078584 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-775fd649b7-h6bmr"] Apr 21 14:34:55.079120 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.078935 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" containerName="console" Apr 21 14:34:55.079120 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.078949 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" containerName="console" Apr 21 14:34:55.079120 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.078990 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a3b5fe8-b6c0-4431-bddc-e4f190739eb6" containerName="console" Apr 21 14:34:55.081862 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.081837 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.099012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.098985 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-775fd649b7-h6bmr"] Apr 21 14:34:55.176528 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.176484 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-oauth-serving-cert\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.176762 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.176540 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-console-config\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.176762 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.176583 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58fc1059-03d3-4155-a810-ffd876f1e46f-console-serving-cert\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.176762 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.176649 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd454\" (UniqueName: \"kubernetes.io/projected/58fc1059-03d3-4155-a810-ffd876f1e46f-kube-api-access-pd454\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.176762 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.176714 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58fc1059-03d3-4155-a810-ffd876f1e46f-console-oauth-config\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.176762 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.176739 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-trusted-ca-bundle\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.176949 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.176785 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-service-ca\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.277916 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.277879 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-oauth-serving-cert\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.277916 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.277920 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-console-config\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.278182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.277939 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58fc1059-03d3-4155-a810-ffd876f1e46f-console-serving-cert\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.278182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.277955 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pd454\" (UniqueName: \"kubernetes.io/projected/58fc1059-03d3-4155-a810-ffd876f1e46f-kube-api-access-pd454\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.278182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.277981 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58fc1059-03d3-4155-a810-ffd876f1e46f-console-oauth-config\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.278182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.277996 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-trusted-ca-bundle\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.278182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.278033 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-service-ca\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.278784 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.278756 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-oauth-serving-cert\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.279016 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.278986 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-service-ca\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.279134 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.278993 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-console-config\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.279134 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.279029 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58fc1059-03d3-4155-a810-ffd876f1e46f-trusted-ca-bundle\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.280655 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.280635 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/58fc1059-03d3-4155-a810-ffd876f1e46f-console-oauth-config\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.280851 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.280811 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/58fc1059-03d3-4155-a810-ffd876f1e46f-console-serving-cert\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.285943 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.285918 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd454\" (UniqueName: \"kubernetes.io/projected/58fc1059-03d3-4155-a810-ffd876f1e46f-kube-api-access-pd454\") pod \"console-775fd649b7-h6bmr\" (UID: \"58fc1059-03d3-4155-a810-ffd876f1e46f\") " pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.391664 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.391556 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:34:55.522770 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.522738 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-775fd649b7-h6bmr"] Apr 21 14:34:55.527088 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:34:55.527062 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58fc1059_03d3_4155_a810_ffd876f1e46f.slice/crio-6c95cad8856872b1c52ce32337af7c4fe554072592d97bae62d4a343ab6562a5 WatchSource:0}: Error finding container 6c95cad8856872b1c52ce32337af7c4fe554072592d97bae62d4a343ab6562a5: Status 404 returned error can't find the container with id 6c95cad8856872b1c52ce32337af7c4fe554072592d97bae62d4a343ab6562a5 Apr 21 14:34:55.925313 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.925271 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-775fd649b7-h6bmr" event={"ID":"58fc1059-03d3-4155-a810-ffd876f1e46f","Type":"ContainerStarted","Data":"8cff5e86fed44a35ed211b2adb10a22f17822a54410aece0249ba26590df3bd4"} Apr 21 14:34:55.925476 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.925320 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-775fd649b7-h6bmr" event={"ID":"58fc1059-03d3-4155-a810-ffd876f1e46f","Type":"ContainerStarted","Data":"6c95cad8856872b1c52ce32337af7c4fe554072592d97bae62d4a343ab6562a5"} Apr 21 14:34:55.943903 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:34:55.943847 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-775fd649b7-h6bmr" podStartSLOduration=0.943832007 podStartE2EDuration="943.832007ms" podCreationTimestamp="2026-04-21 14:34:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:34:55.942677057 +0000 UTC m=+529.400161586" watchObservedRunningTime="2026-04-21 14:34:55.943832007 +0000 UTC m=+529.401316514" Apr 21 14:35:05.392495 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:05.392399 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:35:05.393019 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:05.392531 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:35:05.397479 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:05.397454 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:35:05.957633 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:05.957608 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-775fd649b7-h6bmr" Apr 21 14:35:06.003705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:06.003667 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-dd4f9cd68-nknr5"] Apr 21 14:35:31.025225 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.025179 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-dd4f9cd68-nknr5" podUID="12116f2b-7c82-4ce8-981d-d10e7fda1763" containerName="console" containerID="cri-o://ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222" gracePeriod=15 Apr 21 14:35:31.267759 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.267734 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-dd4f9cd68-nknr5_12116f2b-7c82-4ce8-981d-d10e7fda1763/console/0.log" Apr 21 14:35:31.267911 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.267800 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:35:31.367142 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367046 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-config\") pod \"12116f2b-7c82-4ce8-981d-d10e7fda1763\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " Apr 21 14:35:31.367142 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367119 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-oauth-config\") pod \"12116f2b-7c82-4ce8-981d-d10e7fda1763\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " Apr 21 14:35:31.367366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367151 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-service-ca\") pod \"12116f2b-7c82-4ce8-981d-d10e7fda1763\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " Apr 21 14:35:31.367366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367176 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjslb\" (UniqueName: \"kubernetes.io/projected/12116f2b-7c82-4ce8-981d-d10e7fda1763-kube-api-access-wjslb\") pod \"12116f2b-7c82-4ce8-981d-d10e7fda1763\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " Apr 21 14:35:31.367366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367225 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-oauth-serving-cert\") pod \"12116f2b-7c82-4ce8-981d-d10e7fda1763\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " Apr 21 14:35:31.367366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367248 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-serving-cert\") pod \"12116f2b-7c82-4ce8-981d-d10e7fda1763\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " Apr 21 14:35:31.367366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367302 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-trusted-ca-bundle\") pod \"12116f2b-7c82-4ce8-981d-d10e7fda1763\" (UID: \"12116f2b-7c82-4ce8-981d-d10e7fda1763\") " Apr 21 14:35:31.367366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367316 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-config" (OuterVolumeSpecName: "console-config") pod "12116f2b-7c82-4ce8-981d-d10e7fda1763" (UID: "12116f2b-7c82-4ce8-981d-d10e7fda1763"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:35:31.367684 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367536 2580 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:35:31.367746 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367597 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-service-ca" (OuterVolumeSpecName: "service-ca") pod "12116f2b-7c82-4ce8-981d-d10e7fda1763" (UID: "12116f2b-7c82-4ce8-981d-d10e7fda1763"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:35:31.367746 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367720 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "12116f2b-7c82-4ce8-981d-d10e7fda1763" (UID: "12116f2b-7c82-4ce8-981d-d10e7fda1763"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:35:31.367847 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.367769 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "12116f2b-7c82-4ce8-981d-d10e7fda1763" (UID: "12116f2b-7c82-4ce8-981d-d10e7fda1763"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:35:31.369610 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.369583 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "12116f2b-7c82-4ce8-981d-d10e7fda1763" (UID: "12116f2b-7c82-4ce8-981d-d10e7fda1763"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:35:31.369723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.369651 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "12116f2b-7c82-4ce8-981d-d10e7fda1763" (UID: "12116f2b-7c82-4ce8-981d-d10e7fda1763"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:35:31.369723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.369661 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12116f2b-7c82-4ce8-981d-d10e7fda1763-kube-api-access-wjslb" (OuterVolumeSpecName: "kube-api-access-wjslb") pod "12116f2b-7c82-4ce8-981d-d10e7fda1763" (UID: "12116f2b-7c82-4ce8-981d-d10e7fda1763"). InnerVolumeSpecName "kube-api-access-wjslb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:35:31.468164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.468126 2580 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-service-ca\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:35:31.468164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.468159 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wjslb\" (UniqueName: \"kubernetes.io/projected/12116f2b-7c82-4ce8-981d-d10e7fda1763-kube-api-access-wjslb\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:35:31.468164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.468171 2580 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-oauth-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:35:31.468390 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.468180 2580 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-serving-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:35:31.468390 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.468188 2580 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12116f2b-7c82-4ce8-981d-d10e7fda1763-trusted-ca-bundle\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:35:31.468390 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:31.468197 2580 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12116f2b-7c82-4ce8-981d-d10e7fda1763-console-oauth-config\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:35:32.027785 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.027756 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-dd4f9cd68-nknr5_12116f2b-7c82-4ce8-981d-d10e7fda1763/console/0.log" Apr 21 14:35:32.028199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.027796 2580 generic.go:358] "Generic (PLEG): container finished" podID="12116f2b-7c82-4ce8-981d-d10e7fda1763" containerID="ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222" exitCode=2 Apr 21 14:35:32.028199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.027832 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dd4f9cd68-nknr5" event={"ID":"12116f2b-7c82-4ce8-981d-d10e7fda1763","Type":"ContainerDied","Data":"ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222"} Apr 21 14:35:32.028199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.027870 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dd4f9cd68-nknr5" Apr 21 14:35:32.028199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.027884 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dd4f9cd68-nknr5" event={"ID":"12116f2b-7c82-4ce8-981d-d10e7fda1763","Type":"ContainerDied","Data":"96ff4bb24bf449b0753dfa25cdebbd2ee9a5bcb3b99de201f21823f83b874188"} Apr 21 14:35:32.028199 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.027904 2580 scope.go:117] "RemoveContainer" containerID="ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222" Apr 21 14:35:32.038094 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.038069 2580 scope.go:117] "RemoveContainer" containerID="ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222" Apr 21 14:35:32.038673 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:35:32.038644 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222\": container with ID starting with ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222 not found: ID does not exist" containerID="ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222" Apr 21 14:35:32.038765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.038687 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222"} err="failed to get container status \"ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222\": rpc error: code = NotFound desc = could not find container \"ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222\": container with ID starting with ff735eb847132609e66940d54e83a2bba6ece4399a0c0bb33f91a459ae53d222 not found: ID does not exist" Apr 21 14:35:32.060147 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.060108 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-dd4f9cd68-nknr5"] Apr 21 14:35:32.066831 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:32.066806 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-dd4f9cd68-nknr5"] Apr 21 14:35:33.245941 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:35:33.245903 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12116f2b-7c82-4ce8-981d-d10e7fda1763" path="/var/lib/kubelet/pods/12116f2b-7c82-4ce8-981d-d10e7fda1763/volumes" Apr 21 14:36:07.129999 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:07.129973 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:36:07.131588 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:07.131550 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:36:07.131846 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:07.131825 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:36:07.133650 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:07.133627 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:36:07.135015 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:07.134991 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:36:07.136823 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:07.136799 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:36:17.761232 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.761195 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-9r7mg"] Apr 21 14:36:17.761658 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.761464 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="12116f2b-7c82-4ce8-981d-d10e7fda1763" containerName="console" Apr 21 14:36:17.761658 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.761474 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="12116f2b-7c82-4ce8-981d-d10e7fda1763" containerName="console" Apr 21 14:36:17.761658 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.761528 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="12116f2b-7c82-4ce8-981d-d10e7fda1763" containerName="console" Apr 21 14:36:17.765492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.765475 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:17.768193 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.768171 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Apr 21 14:36:17.768299 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.768211 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-pc5ft\"" Apr 21 14:36:17.768299 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.768214 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Apr 21 14:36:17.776752 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.776729 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-9r7mg"] Apr 21 14:36:17.845531 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.845493 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbldk\" (UniqueName: \"kubernetes.io/projected/e033dfa7-a8bb-4b8f-b827-91c0bddbabe1-kube-api-access-mbldk\") pod \"cert-manager-webhook-587ccfb98-9r7mg\" (UID: \"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1\") " pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:17.845531 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.845542 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e033dfa7-a8bb-4b8f-b827-91c0bddbabe1-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-9r7mg\" (UID: \"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1\") " pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:17.946770 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.946728 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mbldk\" (UniqueName: \"kubernetes.io/projected/e033dfa7-a8bb-4b8f-b827-91c0bddbabe1-kube-api-access-mbldk\") pod \"cert-manager-webhook-587ccfb98-9r7mg\" (UID: \"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1\") " pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:17.946970 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.946779 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e033dfa7-a8bb-4b8f-b827-91c0bddbabe1-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-9r7mg\" (UID: \"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1\") " pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:17.969902 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.969872 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e033dfa7-a8bb-4b8f-b827-91c0bddbabe1-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-9r7mg\" (UID: \"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1\") " pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:17.974057 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:17.974030 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbldk\" (UniqueName: \"kubernetes.io/projected/e033dfa7-a8bb-4b8f-b827-91c0bddbabe1-kube-api-access-mbldk\") pod \"cert-manager-webhook-587ccfb98-9r7mg\" (UID: \"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1\") " pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:18.091181 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:18.091100 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:18.223796 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:18.223765 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-9r7mg"] Apr 21 14:36:18.226825 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:36:18.226791 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode033dfa7_a8bb_4b8f_b827_91c0bddbabe1.slice/crio-3dd1bca8be650431b34ea5e4d9f153616ba30d93283d15a8b337fb499d8b1814 WatchSource:0}: Error finding container 3dd1bca8be650431b34ea5e4d9f153616ba30d93283d15a8b337fb499d8b1814: Status 404 returned error can't find the container with id 3dd1bca8be650431b34ea5e4d9f153616ba30d93283d15a8b337fb499d8b1814 Apr 21 14:36:19.160836 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:19.160799 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" event={"ID":"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1","Type":"ContainerStarted","Data":"3dd1bca8be650431b34ea5e4d9f153616ba30d93283d15a8b337fb499d8b1814"} Apr 21 14:36:22.173318 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:22.173279 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" event={"ID":"e033dfa7-a8bb-4b8f-b827-91c0bddbabe1","Type":"ContainerStarted","Data":"0d1127ee4d89f996c67acad3626190a67479ccb42150da5395295cbb80a39a40"} Apr 21 14:36:22.173728 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:22.173393 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:22.202406 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:22.202359 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" podStartSLOduration=2.105730693 podStartE2EDuration="5.202346125s" podCreationTimestamp="2026-04-21 14:36:17 +0000 UTC" firstStartedPulling="2026-04-21 14:36:18.228697575 +0000 UTC m=+611.686182060" lastFinishedPulling="2026-04-21 14:36:21.325313003 +0000 UTC m=+614.782797492" observedRunningTime="2026-04-21 14:36:22.20171787 +0000 UTC m=+615.659202377" watchObservedRunningTime="2026-04-21 14:36:22.202346125 +0000 UTC m=+615.659830632" Apr 21 14:36:28.179444 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:28.179406 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-587ccfb98-9r7mg" Apr 21 14:36:44.827640 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.827601 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w"] Apr 21 14:36:44.830877 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.830857 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:44.833545 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.833524 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"openshift-service-ca.crt\"" Apr 21 14:36:44.834269 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.834249 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"webhook-server-cert\"" Apr 21 14:36:44.834390 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.834283 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"lws-manager-config\"" Apr 21 14:36:44.834390 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.834255 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"kube-root-ca.crt\"" Apr 21 14:36:44.834499 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.834408 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"metrics-server-cert\"" Apr 21 14:36:44.834729 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.834712 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"lws-controller-manager-dockercfg-tzj2r\"" Apr 21 14:36:44.841598 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.841577 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w"] Apr 21 14:36:44.960755 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.960714 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5hnv\" (UniqueName: \"kubernetes.io/projected/f445aa9d-90a6-44fa-9045-078ddba5fc21-kube-api-access-b5hnv\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:44.960956 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.960779 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/f445aa9d-90a6-44fa-9045-078ddba5fc21-metrics-cert\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:44.960956 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.960847 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f445aa9d-90a6-44fa-9045-078ddba5fc21-manager-config\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:44.960956 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:44.960911 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f445aa9d-90a6-44fa-9045-078ddba5fc21-cert\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.062131 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.062077 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f445aa9d-90a6-44fa-9045-078ddba5fc21-cert\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.062131 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.062144 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b5hnv\" (UniqueName: \"kubernetes.io/projected/f445aa9d-90a6-44fa-9045-078ddba5fc21-kube-api-access-b5hnv\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.062361 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.062183 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/f445aa9d-90a6-44fa-9045-078ddba5fc21-metrics-cert\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.062361 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.062201 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f445aa9d-90a6-44fa-9045-078ddba5fc21-manager-config\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.062939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.062918 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f445aa9d-90a6-44fa-9045-078ddba5fc21-manager-config\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.064802 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.064780 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f445aa9d-90a6-44fa-9045-078ddba5fc21-cert\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.064873 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.064783 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/f445aa9d-90a6-44fa-9045-078ddba5fc21-metrics-cert\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.070755 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.070712 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5hnv\" (UniqueName: \"kubernetes.io/projected/f445aa9d-90a6-44fa-9045-078ddba5fc21-kube-api-access-b5hnv\") pod \"lws-controller-manager-55dd4758fb-fkh4w\" (UID: \"f445aa9d-90a6-44fa-9045-078ddba5fc21\") " pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.140849 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.140778 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:45.275143 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.275110 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w"] Apr 21 14:36:45.278181 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:36:45.278143 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf445aa9d_90a6_44fa_9045_078ddba5fc21.slice/crio-051855c13a30ddc5c72066f26c4387e4418c00faa835502ca6cc7eb6a01f607e WatchSource:0}: Error finding container 051855c13a30ddc5c72066f26c4387e4418c00faa835502ca6cc7eb6a01f607e: Status 404 returned error can't find the container with id 051855c13a30ddc5c72066f26c4387e4418c00faa835502ca6cc7eb6a01f607e Apr 21 14:36:45.279719 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:45.279702 2580 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 14:36:46.245188 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:46.245154 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" event={"ID":"f445aa9d-90a6-44fa-9045-078ddba5fc21","Type":"ContainerStarted","Data":"051855c13a30ddc5c72066f26c4387e4418c00faa835502ca6cc7eb6a01f607e"} Apr 21 14:36:48.252021 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:48.251928 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" event={"ID":"f445aa9d-90a6-44fa-9045-078ddba5fc21","Type":"ContainerStarted","Data":"8171ab9bc6cdcd9331514f34d118adfaf2029c2a6f479b5674870ce49639efcb"} Apr 21 14:36:48.252385 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:48.252071 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:36:48.286297 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:48.286243 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" podStartSLOduration=1.581727705 podStartE2EDuration="4.286227704s" podCreationTimestamp="2026-04-21 14:36:44 +0000 UTC" firstStartedPulling="2026-04-21 14:36:45.279824513 +0000 UTC m=+638.737308999" lastFinishedPulling="2026-04-21 14:36:47.984324513 +0000 UTC m=+641.441808998" observedRunningTime="2026-04-21 14:36:48.28547318 +0000 UTC m=+641.742957688" watchObservedRunningTime="2026-04-21 14:36:48.286227704 +0000 UTC m=+641.743712211" Apr 21 14:36:50.763948 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.763910 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk"] Apr 21 14:36:50.767278 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.767258 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.771253 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.771228 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-webhook-cert\"" Apr 21 14:36:50.771397 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.771227 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-manager-service-cert\"" Apr 21 14:36:50.771397 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.771363 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-manager-dockercfg-dt2r5\"" Apr 21 14:36:50.772186 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.772165 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"kube-root-ca.crt\"" Apr 21 14:36:50.776190 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.776174 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"openshift-service-ca.crt\"" Apr 21 14:36:50.801888 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.801857 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk"] Apr 21 14:36:50.809425 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.809395 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-webhook-cert\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.809558 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.809438 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-apiservice-cert\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.809558 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.809461 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m54qh\" (UniqueName: \"kubernetes.io/projected/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-kube-api-access-m54qh\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.910693 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.910656 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-webhook-cert\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.910877 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.910701 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-apiservice-cert\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.910877 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.910741 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m54qh\" (UniqueName: \"kubernetes.io/projected/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-kube-api-access-m54qh\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.913425 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.913390 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-apiservice-cert\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.913528 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.913393 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-webhook-cert\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:50.934173 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:50.934138 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m54qh\" (UniqueName: \"kubernetes.io/projected/fbebfd28-f8f1-40d8-83f8-99bbbb82a630-kube-api-access-m54qh\") pod \"opendatahub-operator-controller-manager-7df645bd74-9mxmk\" (UID: \"fbebfd28-f8f1-40d8-83f8-99bbbb82a630\") " pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:51.077507 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:51.077408 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:51.209511 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:51.209475 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk"] Apr 21 14:36:51.212669 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:36:51.212640 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbebfd28_f8f1_40d8_83f8_99bbbb82a630.slice/crio-d293e330b684d0d0d93d57811860de97b42d3b8f4bbd4b28aff62054cd4c0ac6 WatchSource:0}: Error finding container d293e330b684d0d0d93d57811860de97b42d3b8f4bbd4b28aff62054cd4c0ac6: Status 404 returned error can't find the container with id d293e330b684d0d0d93d57811860de97b42d3b8f4bbd4b28aff62054cd4c0ac6 Apr 21 14:36:51.260861 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:51.260823 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" event={"ID":"fbebfd28-f8f1-40d8-83f8-99bbbb82a630","Type":"ContainerStarted","Data":"d293e330b684d0d0d93d57811860de97b42d3b8f4bbd4b28aff62054cd4c0ac6"} Apr 21 14:36:55.274940 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:55.274901 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" event={"ID":"fbebfd28-f8f1-40d8-83f8-99bbbb82a630","Type":"ContainerStarted","Data":"8db32f88263a1ac297ee16aac12f1ae188157e6e4ed436044a4908fa04e127e7"} Apr 21 14:36:55.275395 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:55.275053 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:36:55.298326 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:55.298271 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" podStartSLOduration=2.193936663 podStartE2EDuration="5.298254524s" podCreationTimestamp="2026-04-21 14:36:50 +0000 UTC" firstStartedPulling="2026-04-21 14:36:51.214319109 +0000 UTC m=+644.671803593" lastFinishedPulling="2026-04-21 14:36:54.318636965 +0000 UTC m=+647.776121454" observedRunningTime="2026-04-21 14:36:55.295717132 +0000 UTC m=+648.753201652" watchObservedRunningTime="2026-04-21 14:36:55.298254524 +0000 UTC m=+648.755739032" Apr 21 14:36:59.257182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:36:59.257145 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-lws-operator/lws-controller-manager-55dd4758fb-fkh4w" Apr 21 14:37:06.280340 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:06.280309 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/opendatahub-operator-controller-manager-7df645bd74-9mxmk" Apr 21 14:37:44.641504 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.641471 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d"] Apr 21 14:37:44.651811 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.651787 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.654893 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.654855 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"data-science-gateway-data-science-gateway-class-dockercfg-rq58x\"" Apr 21 14:37:44.655040 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.654898 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"istio-ca-root-cert\"" Apr 21 14:37:44.658484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.658437 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d"] Apr 21 14:37:44.739031 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.738990 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-token\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739257 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739049 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-podinfo\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739257 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739144 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-workload-certs\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739257 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739185 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-envoy\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739257 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-workload-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739287 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-credential-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739322 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlwpj\" (UniqueName: \"kubernetes.io/projected/37d51f19-e7f6-4871-8cb3-289134a6eabb-kube-api-access-qlwpj\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739369 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-data\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.739542 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.739404 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/37d51f19-e7f6-4871-8cb3-289134a6eabb-istiod-ca-cert\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840403 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840359 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-workload-certs\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840403 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840404 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-envoy\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840731 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840436 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-workload-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840731 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840467 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-credential-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840731 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840487 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qlwpj\" (UniqueName: \"kubernetes.io/projected/37d51f19-e7f6-4871-8cb3-289134a6eabb-kube-api-access-qlwpj\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840731 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840519 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-data\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840731 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840549 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/37d51f19-e7f6-4871-8cb3-289134a6eabb-istiod-ca-cert\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840840 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-workload-certs\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840892 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-token\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840961 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-podinfo\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.840985 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.840971 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-workload-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.841192 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.841106 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-data\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.841192 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.841129 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-credential-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.841297 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.841226 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/37d51f19-e7f6-4871-8cb3-289134a6eabb-istiod-ca-cert\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.843165 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.843137 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-envoy\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.843914 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.843893 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-podinfo\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.849775 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.849750 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/37d51f19-e7f6-4871-8cb3-289134a6eabb-istio-token\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.849954 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.849932 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlwpj\" (UniqueName: \"kubernetes.io/projected/37d51f19-e7f6-4871-8cb3-289134a6eabb-kube-api-access-qlwpj\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d\" (UID: \"37d51f19-e7f6-4871-8cb3-289134a6eabb\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:44.964015 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:44.963981 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:45.091764 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:45.091735 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d"] Apr 21 14:37:45.094254 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:37:45.094219 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37d51f19_e7f6_4871_8cb3_289134a6eabb.slice/crio-589f8a941c92613a9d3e3a55abf364014c90e69ed193fed5ed42b968f1b40575 WatchSource:0}: Error finding container 589f8a941c92613a9d3e3a55abf364014c90e69ed193fed5ed42b968f1b40575: Status 404 returned error can't find the container with id 589f8a941c92613a9d3e3a55abf364014c90e69ed193fed5ed42b968f1b40575 Apr 21 14:37:45.433283 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:45.433240 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" event={"ID":"37d51f19-e7f6-4871-8cb3-289134a6eabb","Type":"ContainerStarted","Data":"589f8a941c92613a9d3e3a55abf364014c90e69ed193fed5ed42b968f1b40575"} Apr 21 14:37:47.574580 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:47.574509 2580 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892152Ki","pods":"250"} Apr 21 14:37:47.574953 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:47.574636 2580 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892152Ki","pods":"250"} Apr 21 14:37:47.574953 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:47.574688 2580 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892152Ki","pods":"250"} Apr 21 14:37:48.445856 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:48.445809 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" event={"ID":"37d51f19-e7f6-4871-8cb3-289134a6eabb","Type":"ContainerStarted","Data":"012aafaef4126873799cd4b758fd4ddaa70e342f7ffc8b587cab80303dbe5c68"} Apr 21 14:37:48.466189 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:48.466136 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" podStartSLOduration=1.988192362 podStartE2EDuration="4.466120963s" podCreationTimestamp="2026-04-21 14:37:44 +0000 UTC" firstStartedPulling="2026-04-21 14:37:45.096328471 +0000 UTC m=+698.553812957" lastFinishedPulling="2026-04-21 14:37:47.57425707 +0000 UTC m=+701.031741558" observedRunningTime="2026-04-21 14:37:48.465058521 +0000 UTC m=+701.922543028" watchObservedRunningTime="2026-04-21 14:37:48.466120963 +0000 UTC m=+701.923605470" Apr 21 14:37:48.964155 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:48.964121 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:48.968802 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:48.968777 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:49.449014 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:49.448982 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:37:49.449966 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:37:49.449950 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d" Apr 21 14:38:09.268932 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.268896 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-bhfqq"] Apr 21 14:38:09.274166 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.274148 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" Apr 21 14:38:09.276774 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.276750 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"openshift-service-ca.crt\"" Apr 21 14:38:09.276897 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.276794 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"kube-root-ca.crt\"" Apr 21 14:38:09.277921 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.277811 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"kuadrant-operator-catalog-dockercfg-2hjgq\"" Apr 21 14:38:09.283775 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.283750 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-bhfqq"] Apr 21 14:38:09.351701 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.351663 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjhjm\" (UniqueName: \"kubernetes.io/projected/dc2b855e-71df-4625-9f3f-209ee72805a7-kube-api-access-xjhjm\") pod \"kuadrant-operator-catalog-bhfqq\" (UID: \"dc2b855e-71df-4625-9f3f-209ee72805a7\") " pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" Apr 21 14:38:09.452685 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.452643 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xjhjm\" (UniqueName: \"kubernetes.io/projected/dc2b855e-71df-4625-9f3f-209ee72805a7-kube-api-access-xjhjm\") pod \"kuadrant-operator-catalog-bhfqq\" (UID: \"dc2b855e-71df-4625-9f3f-209ee72805a7\") " pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" Apr 21 14:38:09.461634 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.461606 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjhjm\" (UniqueName: \"kubernetes.io/projected/dc2b855e-71df-4625-9f3f-209ee72805a7-kube-api-access-xjhjm\") pod \"kuadrant-operator-catalog-bhfqq\" (UID: \"dc2b855e-71df-4625-9f3f-209ee72805a7\") " pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" Apr 21 14:38:09.586425 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.586332 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" Apr 21 14:38:09.630291 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.630248 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-bhfqq"] Apr 21 14:38:09.715290 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.715268 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-bhfqq"] Apr 21 14:38:09.717060 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:38:09.717031 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc2b855e_71df_4625_9f3f_209ee72805a7.slice/crio-55cfa494639a5393c030093961994e64f0f993e2a8d1bc4a2fdc92a035a49883 WatchSource:0}: Error finding container 55cfa494639a5393c030093961994e64f0f993e2a8d1bc4a2fdc92a035a49883: Status 404 returned error can't find the container with id 55cfa494639a5393c030093961994e64f0f993e2a8d1bc4a2fdc92a035a49883 Apr 21 14:38:09.846998 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.846915 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-6q598"] Apr 21 14:38:09.851725 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.851708 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:09.863354 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.863328 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-6q598"] Apr 21 14:38:09.956349 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:09.956308 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qrpc\" (UniqueName: \"kubernetes.io/projected/856e4d1a-8dc6-487a-848b-2f081de3c7ff-kube-api-access-9qrpc\") pod \"kuadrant-operator-catalog-6q598\" (UID: \"856e4d1a-8dc6-487a-848b-2f081de3c7ff\") " pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:10.057300 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:10.057271 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9qrpc\" (UniqueName: \"kubernetes.io/projected/856e4d1a-8dc6-487a-848b-2f081de3c7ff-kube-api-access-9qrpc\") pod \"kuadrant-operator-catalog-6q598\" (UID: \"856e4d1a-8dc6-487a-848b-2f081de3c7ff\") " pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:10.066052 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:10.066022 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qrpc\" (UniqueName: \"kubernetes.io/projected/856e4d1a-8dc6-487a-848b-2f081de3c7ff-kube-api-access-9qrpc\") pod \"kuadrant-operator-catalog-6q598\" (UID: \"856e4d1a-8dc6-487a-848b-2f081de3c7ff\") " pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:10.161234 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:10.161147 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:10.293544 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:10.293514 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-6q598"] Apr 21 14:38:10.299760 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:38:10.299733 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod856e4d1a_8dc6_487a_848b_2f081de3c7ff.slice/crio-3e2786c0c042cac5731d5abae5f7452bdd1a5c4004f466d64005444e4a0cf7da WatchSource:0}: Error finding container 3e2786c0c042cac5731d5abae5f7452bdd1a5c4004f466d64005444e4a0cf7da: Status 404 returned error can't find the container with id 3e2786c0c042cac5731d5abae5f7452bdd1a5c4004f466d64005444e4a0cf7da Apr 21 14:38:10.524453 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:10.524400 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-6q598" event={"ID":"856e4d1a-8dc6-487a-848b-2f081de3c7ff","Type":"ContainerStarted","Data":"3e2786c0c042cac5731d5abae5f7452bdd1a5c4004f466d64005444e4a0cf7da"} Apr 21 14:38:10.525555 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:10.525528 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" event={"ID":"dc2b855e-71df-4625-9f3f-209ee72805a7","Type":"ContainerStarted","Data":"55cfa494639a5393c030093961994e64f0f993e2a8d1bc4a2fdc92a035a49883"} Apr 21 14:38:12.534819 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.534782 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" event={"ID":"dc2b855e-71df-4625-9f3f-209ee72805a7","Type":"ContainerStarted","Data":"46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469"} Apr 21 14:38:12.535291 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.534842 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" podUID="dc2b855e-71df-4625-9f3f-209ee72805a7" containerName="registry-server" containerID="cri-o://46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469" gracePeriod=2 Apr 21 14:38:12.536098 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.536063 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-6q598" event={"ID":"856e4d1a-8dc6-487a-848b-2f081de3c7ff","Type":"ContainerStarted","Data":"3b594cf70cee6d4f5c7792de2927c593b1297e2413756ad8e7167339e38d436c"} Apr 21 14:38:12.551839 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.551796 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" podStartSLOduration=1.369586207 podStartE2EDuration="3.551781297s" podCreationTimestamp="2026-04-21 14:38:09 +0000 UTC" firstStartedPulling="2026-04-21 14:38:09.71827458 +0000 UTC m=+723.175759065" lastFinishedPulling="2026-04-21 14:38:11.90046966 +0000 UTC m=+725.357954155" observedRunningTime="2026-04-21 14:38:12.550105747 +0000 UTC m=+726.007590253" watchObservedRunningTime="2026-04-21 14:38:12.551781297 +0000 UTC m=+726.009265804" Apr 21 14:38:12.567893 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.567847 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/kuadrant-operator-catalog-6q598" podStartSLOduration=1.967699671 podStartE2EDuration="3.567833207s" podCreationTimestamp="2026-04-21 14:38:09 +0000 UTC" firstStartedPulling="2026-04-21 14:38:10.301143437 +0000 UTC m=+723.758627925" lastFinishedPulling="2026-04-21 14:38:11.901276973 +0000 UTC m=+725.358761461" observedRunningTime="2026-04-21 14:38:12.566266764 +0000 UTC m=+726.023751271" watchObservedRunningTime="2026-04-21 14:38:12.567833207 +0000 UTC m=+726.025317738" Apr 21 14:38:12.776064 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.776038 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" Apr 21 14:38:12.884304 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.884217 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjhjm\" (UniqueName: \"kubernetes.io/projected/dc2b855e-71df-4625-9f3f-209ee72805a7-kube-api-access-xjhjm\") pod \"dc2b855e-71df-4625-9f3f-209ee72805a7\" (UID: \"dc2b855e-71df-4625-9f3f-209ee72805a7\") " Apr 21 14:38:12.886512 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.886488 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2b855e-71df-4625-9f3f-209ee72805a7-kube-api-access-xjhjm" (OuterVolumeSpecName: "kube-api-access-xjhjm") pod "dc2b855e-71df-4625-9f3f-209ee72805a7" (UID: "dc2b855e-71df-4625-9f3f-209ee72805a7"). InnerVolumeSpecName "kube-api-access-xjhjm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:38:12.985356 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:12.985317 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xjhjm\" (UniqueName: \"kubernetes.io/projected/dc2b855e-71df-4625-9f3f-209ee72805a7-kube-api-access-xjhjm\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:38:13.540318 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.540229 2580 generic.go:358] "Generic (PLEG): container finished" podID="dc2b855e-71df-4625-9f3f-209ee72805a7" containerID="46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469" exitCode=0 Apr 21 14:38:13.540318 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.540291 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" Apr 21 14:38:13.540824 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.540324 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" event={"ID":"dc2b855e-71df-4625-9f3f-209ee72805a7","Type":"ContainerDied","Data":"46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469"} Apr 21 14:38:13.540824 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.540370 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-bhfqq" event={"ID":"dc2b855e-71df-4625-9f3f-209ee72805a7","Type":"ContainerDied","Data":"55cfa494639a5393c030093961994e64f0f993e2a8d1bc4a2fdc92a035a49883"} Apr 21 14:38:13.540824 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.540391 2580 scope.go:117] "RemoveContainer" containerID="46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469" Apr 21 14:38:13.548730 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.548702 2580 scope.go:117] "RemoveContainer" containerID="46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469" Apr 21 14:38:13.548970 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:38:13.548952 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469\": container with ID starting with 46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469 not found: ID does not exist" containerID="46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469" Apr 21 14:38:13.549036 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.548984 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469"} err="failed to get container status \"46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469\": rpc error: code = NotFound desc = could not find container \"46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469\": container with ID starting with 46273692a88797761f15dc456347e1f2eee268fdf8194389ebeaf3e3230fa469 not found: ID does not exist" Apr 21 14:38:13.559751 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.559726 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-bhfqq"] Apr 21 14:38:13.563951 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:13.563927 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-bhfqq"] Apr 21 14:38:15.244937 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:15.244908 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc2b855e-71df-4625-9f3f-209ee72805a7" path="/var/lib/kubelet/pods/dc2b855e-71df-4625-9f3f-209ee72805a7/volumes" Apr 21 14:38:20.161508 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:20.161473 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:20.162112 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:20.161560 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:20.184403 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:20.184370 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:20.585591 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:20.585533 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/kuadrant-operator-catalog-6q598" Apr 21 14:38:37.319699 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.319663 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb"] Apr 21 14:38:37.320073 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.319974 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dc2b855e-71df-4625-9f3f-209ee72805a7" containerName="registry-server" Apr 21 14:38:37.320073 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.319985 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2b855e-71df-4625-9f3f-209ee72805a7" containerName="registry-server" Apr 21 14:38:37.320073 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.320033 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="dc2b855e-71df-4625-9f3f-209ee72805a7" containerName="registry-server" Apr 21 14:38:37.322809 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.322791 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:37.328662 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.328627 2580 status_manager.go:895] "Failed to get status for pod" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" err="pods \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" is forbidden: User \"system:node:ip-10-0-141-61.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-141-61.ec2.internal' and this object" Apr 21 14:38:37.328662 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:38:37.328646 2580 reflector.go:200] "Failed to watch" err="failed to list *v1.Secret: secrets \"kuadrant-operator-controller-manager-dockercfg-75flw\" is forbidden: User \"system:node:ip-10-0-141-61.ec2.internal\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-141-61.ec2.internal' and this object" logger="UnhandledError" reflector="object-\"kuadrant-system\"/\"kuadrant-operator-controller-manager-dockercfg-75flw\"" type="*v1.Secret" Apr 21 14:38:37.370626 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.370595 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb"] Apr 21 14:38:37.382080 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.382050 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsb9b\" (UniqueName: \"kubernetes.io/projected/80b44a2d-5259-432b-a222-22cf11f8358e-kube-api-access-xsb9b\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:37.382229 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.382104 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/80b44a2d-5259-432b-a222-22cf11f8358e-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:37.483353 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.483317 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/80b44a2d-5259-432b-a222-22cf11f8358e-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:37.483518 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.483383 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xsb9b\" (UniqueName: \"kubernetes.io/projected/80b44a2d-5259-432b-a222-22cf11f8358e-kube-api-access-xsb9b\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:37.483823 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.483803 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/80b44a2d-5259-432b-a222-22cf11f8358e-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:37.498259 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:37.498230 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsb9b\" (UniqueName: \"kubernetes.io/projected/80b44a2d-5259-432b-a222-22cf11f8358e-kube-api-access-xsb9b\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:38.155495 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:38.155467 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"kuadrant-operator-controller-manager-dockercfg-75flw\"" Apr 21 14:38:38.163928 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:38.163909 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:38.303185 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:38.303157 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb"] Apr 21 14:38:38.305767 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:38:38.305741 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80b44a2d_5259_432b_a222_22cf11f8358e.slice/crio-48916c8e02d1921e67d89720c7d22d43016c7fb40af173c479af37627a04cff0 WatchSource:0}: Error finding container 48916c8e02d1921e67d89720c7d22d43016c7fb40af173c479af37627a04cff0: Status 404 returned error can't find the container with id 48916c8e02d1921e67d89720c7d22d43016c7fb40af173c479af37627a04cff0 Apr 21 14:38:38.625238 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:38.625186 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" event={"ID":"80b44a2d-5259-432b-a222-22cf11f8358e","Type":"ContainerStarted","Data":"48916c8e02d1921e67d89720c7d22d43016c7fb40af173c479af37627a04cff0"} Apr 21 14:38:39.173074 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.173038 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl"] Apr 21 14:38:39.177739 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.177723 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" Apr 21 14:38:39.180376 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.180357 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"dns-operator-controller-manager-dockercfg-ltct9\"" Apr 21 14:38:39.180443 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.180360 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"dns-operator-controller-env\"" Apr 21 14:38:39.186290 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.186264 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl"] Apr 21 14:38:39.197519 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.197491 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxbnb\" (UniqueName: \"kubernetes.io/projected/95a16468-5acb-42be-a1f1-8ffaf3abf10a-kube-api-access-gxbnb\") pod \"dns-operator-controller-manager-648d5c98bc-xn8rl\" (UID: \"95a16468-5acb-42be-a1f1-8ffaf3abf10a\") " pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" Apr 21 14:38:39.298116 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.298074 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gxbnb\" (UniqueName: \"kubernetes.io/projected/95a16468-5acb-42be-a1f1-8ffaf3abf10a-kube-api-access-gxbnb\") pod \"dns-operator-controller-manager-648d5c98bc-xn8rl\" (UID: \"95a16468-5acb-42be-a1f1-8ffaf3abf10a\") " pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" Apr 21 14:38:39.309763 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.309734 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxbnb\" (UniqueName: \"kubernetes.io/projected/95a16468-5acb-42be-a1f1-8ffaf3abf10a-kube-api-access-gxbnb\") pod \"dns-operator-controller-manager-648d5c98bc-xn8rl\" (UID: \"95a16468-5acb-42be-a1f1-8ffaf3abf10a\") " pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" Apr 21 14:38:39.488706 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.488609 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" Apr 21 14:38:39.632642 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:39.632620 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl"] Apr 21 14:38:39.635267 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:38:39.635235 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95a16468_5acb_42be_a1f1_8ffaf3abf10a.slice/crio-8c36b2b3ec2c83562041c81c4bd9af4d9b69e6850bfca2b54d42aefacef218b1 WatchSource:0}: Error finding container 8c36b2b3ec2c83562041c81c4bd9af4d9b69e6850bfca2b54d42aefacef218b1: Status 404 returned error can't find the container with id 8c36b2b3ec2c83562041c81c4bd9af4d9b69e6850bfca2b54d42aefacef218b1 Apr 21 14:38:40.632945 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:40.632904 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" event={"ID":"95a16468-5acb-42be-a1f1-8ffaf3abf10a","Type":"ContainerStarted","Data":"8c36b2b3ec2c83562041c81c4bd9af4d9b69e6850bfca2b54d42aefacef218b1"} Apr 21 14:38:42.641724 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:42.641688 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" event={"ID":"95a16468-5acb-42be-a1f1-8ffaf3abf10a","Type":"ContainerStarted","Data":"82c006a5acdc6e36fbc67a9f4f7be81607e2c8b8f254a86074034156c71678e3"} Apr 21 14:38:42.642134 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:42.641801 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" Apr 21 14:38:42.686296 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:42.686247 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" podStartSLOduration=1.283571329 podStartE2EDuration="3.686232398s" podCreationTimestamp="2026-04-21 14:38:39 +0000 UTC" firstStartedPulling="2026-04-21 14:38:39.637560147 +0000 UTC m=+753.095044633" lastFinishedPulling="2026-04-21 14:38:42.040221202 +0000 UTC m=+755.497705702" observedRunningTime="2026-04-21 14:38:42.68401099 +0000 UTC m=+756.141495497" watchObservedRunningTime="2026-04-21 14:38:42.686232398 +0000 UTC m=+756.143716904" Apr 21 14:38:44.655501 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:44.655397 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" event={"ID":"80b44a2d-5259-432b-a222-22cf11f8358e","Type":"ContainerStarted","Data":"a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194"} Apr 21 14:38:44.655996 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:44.655523 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:45.679991 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.679942 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" podStartSLOduration=2.709042287 podStartE2EDuration="8.679926531s" podCreationTimestamp="2026-04-21 14:38:37 +0000 UTC" firstStartedPulling="2026-04-21 14:38:38.30809871 +0000 UTC m=+751.765583196" lastFinishedPulling="2026-04-21 14:38:44.278982955 +0000 UTC m=+757.736467440" observedRunningTime="2026-04-21 14:38:44.685026772 +0000 UTC m=+758.142511302" watchObservedRunningTime="2026-04-21 14:38:45.679926531 +0000 UTC m=+759.137411038" Apr 21 14:38:45.680536 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.680519 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-operator-657f44b778-kpwkq"] Apr 21 14:38:45.683808 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.683792 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" Apr 21 14:38:45.686252 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.686228 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-operator-dockercfg-hdc5q\"" Apr 21 14:38:45.703629 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.703604 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-operator-657f44b778-kpwkq"] Apr 21 14:38:45.755129 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.755096 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cr87\" (UniqueName: \"kubernetes.io/projected/dd86a861-3beb-495b-b60e-6361f51519cc-kube-api-access-8cr87\") pod \"authorino-operator-657f44b778-kpwkq\" (UID: \"dd86a861-3beb-495b-b60e-6361f51519cc\") " pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" Apr 21 14:38:45.855971 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.855936 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8cr87\" (UniqueName: \"kubernetes.io/projected/dd86a861-3beb-495b-b60e-6361f51519cc-kube-api-access-8cr87\") pod \"authorino-operator-657f44b778-kpwkq\" (UID: \"dd86a861-3beb-495b-b60e-6361f51519cc\") " pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" Apr 21 14:38:45.869290 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.869258 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cr87\" (UniqueName: \"kubernetes.io/projected/dd86a861-3beb-495b-b60e-6361f51519cc-kube-api-access-8cr87\") pod \"authorino-operator-657f44b778-kpwkq\" (UID: \"dd86a861-3beb-495b-b60e-6361f51519cc\") " pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" Apr 21 14:38:45.993386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:45.993297 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" Apr 21 14:38:46.135255 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:38:46.135226 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd86a861_3beb_495b_b60e_6361f51519cc.slice/crio-1dcf1a373ab0f743e609446d0f4542afa9b8121454fdc4997f88764730a94e59 WatchSource:0}: Error finding container 1dcf1a373ab0f743e609446d0f4542afa9b8121454fdc4997f88764730a94e59: Status 404 returned error can't find the container with id 1dcf1a373ab0f743e609446d0f4542afa9b8121454fdc4997f88764730a94e59 Apr 21 14:38:46.163505 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:46.163468 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-operator-657f44b778-kpwkq"] Apr 21 14:38:46.663685 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:46.663623 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" event={"ID":"dd86a861-3beb-495b-b60e-6361f51519cc","Type":"ContainerStarted","Data":"1dcf1a373ab0f743e609446d0f4542afa9b8121454fdc4997f88764730a94e59"} Apr 21 14:38:48.672420 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:48.672383 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" event={"ID":"dd86a861-3beb-495b-b60e-6361f51519cc","Type":"ContainerStarted","Data":"dc3dfc243ca0f0ee756f6abfe53941b23980708d2766ec6b1ed44c592a8bb885"} Apr 21 14:38:48.672842 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:48.672442 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" Apr 21 14:38:48.693270 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:48.693220 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" podStartSLOduration=1.551009348 podStartE2EDuration="3.693205498s" podCreationTimestamp="2026-04-21 14:38:45 +0000 UTC" firstStartedPulling="2026-04-21 14:38:46.137526886 +0000 UTC m=+759.595011371" lastFinishedPulling="2026-04-21 14:38:48.279723036 +0000 UTC m=+761.737207521" observedRunningTime="2026-04-21 14:38:48.691872334 +0000 UTC m=+762.149356842" watchObservedRunningTime="2026-04-21 14:38:48.693205498 +0000 UTC m=+762.150690004" Apr 21 14:38:53.648687 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:53.648651 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/dns-operator-controller-manager-648d5c98bc-xn8rl" Apr 21 14:38:55.660554 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:55.660514 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:57.287375 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.287344 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb"] Apr 21 14:38:57.287826 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.287557 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" containerName="manager" containerID="cri-o://a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194" gracePeriod=2 Apr 21 14:38:57.300552 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.300509 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb"] Apr 21 14:38:57.317438 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.317414 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4"] Apr 21 14:38:57.317813 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.317797 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" containerName="manager" Apr 21 14:38:57.317910 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.317816 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" containerName="manager" Apr 21 14:38:57.317910 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.317923 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" containerName="manager" Apr 21 14:38:57.320956 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.320937 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.334692 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.334638 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4"] Apr 21 14:38:57.359315 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.359274 2580 status_manager.go:895] "Failed to get status for pod" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" err="pods \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" is forbidden: User \"system:node:ip-10-0-141-61.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-141-61.ec2.internal' and this object" Apr 21 14:38:57.464169 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.464138 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7zgh\" (UniqueName: \"kubernetes.io/projected/c545f539-a944-4247-8140-059d427fce20-kube-api-access-n7zgh\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-wkmj4\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.464298 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.464181 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/c545f539-a944-4247-8140-059d427fce20-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-wkmj4\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.518711 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.518689 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:57.521032 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.521008 2580 status_manager.go:895] "Failed to get status for pod" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" err="pods \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" is forbidden: User \"system:node:ip-10-0-141-61.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-141-61.ec2.internal' and this object" Apr 21 14:38:57.564691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.564624 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n7zgh\" (UniqueName: \"kubernetes.io/projected/c545f539-a944-4247-8140-059d427fce20-kube-api-access-n7zgh\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-wkmj4\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.564691 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.564664 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/c545f539-a944-4247-8140-059d427fce20-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-wkmj4\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.565113 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.565091 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/c545f539-a944-4247-8140-059d427fce20-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-wkmj4\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.580298 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.580279 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7zgh\" (UniqueName: \"kubernetes.io/projected/c545f539-a944-4247-8140-059d427fce20-kube-api-access-n7zgh\") pod \"kuadrant-operator-controller-manager-5f895dd7d5-wkmj4\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.665879 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.665845 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/80b44a2d-5259-432b-a222-22cf11f8358e-extensions-socket-volume\") pod \"80b44a2d-5259-432b-a222-22cf11f8358e\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " Apr 21 14:38:57.666073 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.665916 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsb9b\" (UniqueName: \"kubernetes.io/projected/80b44a2d-5259-432b-a222-22cf11f8358e-kube-api-access-xsb9b\") pod \"80b44a2d-5259-432b-a222-22cf11f8358e\" (UID: \"80b44a2d-5259-432b-a222-22cf11f8358e\") " Apr 21 14:38:57.666365 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.666341 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b44a2d-5259-432b-a222-22cf11f8358e-extensions-socket-volume" (OuterVolumeSpecName: "extensions-socket-volume") pod "80b44a2d-5259-432b-a222-22cf11f8358e" (UID: "80b44a2d-5259-432b-a222-22cf11f8358e"). InnerVolumeSpecName "extensions-socket-volume". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 14:38:57.668155 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.668125 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b44a2d-5259-432b-a222-22cf11f8358e-kube-api-access-xsb9b" (OuterVolumeSpecName: "kube-api-access-xsb9b") pod "80b44a2d-5259-432b-a222-22cf11f8358e" (UID: "80b44a2d-5259-432b-a222-22cf11f8358e"). InnerVolumeSpecName "kube-api-access-xsb9b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:38:57.671294 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.671271 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:57.703843 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.703808 2580 generic.go:358] "Generic (PLEG): container finished" podID="80b44a2d-5259-432b-a222-22cf11f8358e" containerID="a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194" exitCode=0 Apr 21 14:38:57.703989 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.703850 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" Apr 21 14:38:57.703989 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.703876 2580 scope.go:117] "RemoveContainer" containerID="a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194" Apr 21 14:38:57.706114 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.706068 2580 status_manager.go:895] "Failed to get status for pod" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" err="pods \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" is forbidden: User \"system:node:ip-10-0-141-61.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-141-61.ec2.internal' and this object" Apr 21 14:38:57.713373 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.713354 2580 scope.go:117] "RemoveContainer" containerID="a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194" Apr 21 14:38:57.713661 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:38:57.713643 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194\": container with ID starting with a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194 not found: ID does not exist" containerID="a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194" Apr 21 14:38:57.713742 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.713668 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194"} err="failed to get container status \"a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194\": rpc error: code = NotFound desc = could not find container \"a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194\": container with ID starting with a26e7cc4d959b0c375149f62674d8cae832d2fd1c537f715267a79b7d370c194 not found: ID does not exist" Apr 21 14:38:57.715179 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.715160 2580 status_manager.go:895] "Failed to get status for pod" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" err="pods \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" is forbidden: User \"system:node:ip-10-0-141-61.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-141-61.ec2.internal' and this object" Apr 21 14:38:57.767150 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.767116 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xsb9b\" (UniqueName: \"kubernetes.io/projected/80b44a2d-5259-432b-a222-22cf11f8358e-kube-api-access-xsb9b\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:38:57.767150 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.767147 2580 reconciler_common.go:299] "Volume detached for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/80b44a2d-5259-432b-a222-22cf11f8358e-extensions-socket-volume\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:38:57.806207 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:57.806182 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4"] Apr 21 14:38:57.808296 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:38:57.808270 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc545f539_a944_4247_8140_059d427fce20.slice/crio-5dd0a2140763106d685daa7a3955dff3e3213f0b36875a4be3e84e27bc871fb5 WatchSource:0}: Error finding container 5dd0a2140763106d685daa7a3955dff3e3213f0b36875a4be3e84e27bc871fb5: Status 404 returned error can't find the container with id 5dd0a2140763106d685daa7a3955dff3e3213f0b36875a4be3e84e27bc871fb5 Apr 21 14:38:58.708678 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:58.708645 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" event={"ID":"c545f539-a944-4247-8140-059d427fce20","Type":"ContainerStarted","Data":"2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef"} Apr 21 14:38:58.708678 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:58.708680 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" event={"ID":"c545f539-a944-4247-8140-059d427fce20","Type":"ContainerStarted","Data":"5dd0a2140763106d685daa7a3955dff3e3213f0b36875a4be3e84e27bc871fb5"} Apr 21 14:38:58.709111 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:58.708788 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:38:58.710773 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:58.710736 2580 status_manager.go:895] "Failed to get status for pod" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-lqdxb" err="pods \"kuadrant-operator-controller-manager-5f895dd7d5-lqdxb\" is forbidden: User \"system:node:ip-10-0-141-61.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-141-61.ec2.internal' and this object" Apr 21 14:38:58.741249 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:58.741199 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" podStartSLOduration=1.7411857039999998 podStartE2EDuration="1.741185704s" podCreationTimestamp="2026-04-21 14:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:38:58.740147955 +0000 UTC m=+772.197632462" watchObservedRunningTime="2026-04-21 14:38:58.741185704 +0000 UTC m=+772.198670212" Apr 21 14:38:59.246048 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:59.246009 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b44a2d-5259-432b-a222-22cf11f8358e" path="/var/lib/kubelet/pods/80b44a2d-5259-432b-a222-22cf11f8358e/volumes" Apr 21 14:38:59.678024 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:38:59.677996 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/authorino-operator-657f44b778-kpwkq" Apr 21 14:39:09.715193 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:09.715151 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:39:14.293995 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:14.293957 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4"] Apr 21 14:39:14.294445 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:14.294172 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" podUID="c545f539-a944-4247-8140-059d427fce20" containerName="manager" containerID="cri-o://2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef" gracePeriod=10 Apr 21 14:39:14.927850 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:14.927826 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:39:15.019053 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.018957 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7zgh\" (UniqueName: \"kubernetes.io/projected/c545f539-a944-4247-8140-059d427fce20-kube-api-access-n7zgh\") pod \"c545f539-a944-4247-8140-059d427fce20\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " Apr 21 14:39:15.019053 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.019000 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/c545f539-a944-4247-8140-059d427fce20-extensions-socket-volume\") pod \"c545f539-a944-4247-8140-059d427fce20\" (UID: \"c545f539-a944-4247-8140-059d427fce20\") " Apr 21 14:39:15.019398 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.019375 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c545f539-a944-4247-8140-059d427fce20-extensions-socket-volume" (OuterVolumeSpecName: "extensions-socket-volume") pod "c545f539-a944-4247-8140-059d427fce20" (UID: "c545f539-a944-4247-8140-059d427fce20"). InnerVolumeSpecName "extensions-socket-volume". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 14:39:15.021239 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.021212 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c545f539-a944-4247-8140-059d427fce20-kube-api-access-n7zgh" (OuterVolumeSpecName: "kube-api-access-n7zgh") pod "c545f539-a944-4247-8140-059d427fce20" (UID: "c545f539-a944-4247-8140-059d427fce20"). InnerVolumeSpecName "kube-api-access-n7zgh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:39:15.119776 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.119728 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-n7zgh\" (UniqueName: \"kubernetes.io/projected/c545f539-a944-4247-8140-059d427fce20-kube-api-access-n7zgh\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:39:15.119776 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.119770 2580 reconciler_common.go:299] "Volume detached for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/c545f539-a944-4247-8140-059d427fce20-extensions-socket-volume\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:39:15.765427 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.765391 2580 generic.go:358] "Generic (PLEG): container finished" podID="c545f539-a944-4247-8140-059d427fce20" containerID="2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef" exitCode=0 Apr 21 14:39:15.765923 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.765454 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" event={"ID":"c545f539-a944-4247-8140-059d427fce20","Type":"ContainerDied","Data":"2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef"} Apr 21 14:39:15.765923 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.765483 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" event={"ID":"c545f539-a944-4247-8140-059d427fce20","Type":"ContainerDied","Data":"5dd0a2140763106d685daa7a3955dff3e3213f0b36875a4be3e84e27bc871fb5"} Apr 21 14:39:15.765923 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.765481 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4" Apr 21 14:39:15.765923 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.765496 2580 scope.go:117] "RemoveContainer" containerID="2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef" Apr 21 14:39:15.773672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.773641 2580 scope.go:117] "RemoveContainer" containerID="2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef" Apr 21 14:39:15.773910 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:39:15.773894 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef\": container with ID starting with 2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef not found: ID does not exist" containerID="2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef" Apr 21 14:39:15.773956 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.773924 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef"} err="failed to get container status \"2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef\": rpc error: code = NotFound desc = could not find container \"2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef\": container with ID starting with 2d1467f7214ff38fe105f54d8a468a6b91550f69e47fd48b42542e8e1919ccef not found: ID does not exist" Apr 21 14:39:15.786855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.786826 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4"] Apr 21 14:39:15.790740 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:15.790719 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-5f895dd7d5-wkmj4"] Apr 21 14:39:17.245827 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:17.245793 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c545f539-a944-4247-8140-059d427fce20" path="/var/lib/kubelet/pods/c545f539-a944-4247-8140-059d427fce20/volumes" Apr 21 14:39:30.512024 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.511982 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt"] Apr 21 14:39:30.512484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.512362 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c545f539-a944-4247-8140-059d427fce20" containerName="manager" Apr 21 14:39:30.512484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.512374 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="c545f539-a944-4247-8140-059d427fce20" containerName="manager" Apr 21 14:39:30.512484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.512438 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="c545f539-a944-4247-8140-059d427fce20" containerName="manager" Apr 21 14:39:30.515495 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.515475 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.517837 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.517812 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"maas-default-gateway-openshift-default-dockercfg-4s7nx\"" Apr 21 14:39:30.537006 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.536975 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt"] Apr 21 14:39:30.541453 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541429 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-workload-certs\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541590 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541478 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-credential-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541590 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541515 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/9017a28a-4013-4114-b8fb-4d3612657389-istio-token\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541590 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541553 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrx6m\" (UniqueName: \"kubernetes.io/projected/9017a28a-4013-4114-b8fb-4d3612657389-kube-api-access-qrx6m\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541615 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-workload-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541644 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-istio-envoy\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541668 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/9017a28a-4013-4114-b8fb-4d3612657389-istio-podinfo\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541688 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/9017a28a-4013-4114-b8fb-4d3612657389-istiod-ca-cert\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.541723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.541706 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-istio-data\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.642910 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.642875 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qrx6m\" (UniqueName: \"kubernetes.io/projected/9017a28a-4013-4114-b8fb-4d3612657389-kube-api-access-qrx6m\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.642910 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.642916 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-workload-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643181 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.642935 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-istio-envoy\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643181 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643124 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/9017a28a-4013-4114-b8fb-4d3612657389-istio-podinfo\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643181 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643163 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/9017a28a-4013-4114-b8fb-4d3612657389-istiod-ca-cert\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643332 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643192 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-istio-data\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643332 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643251 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-workload-certs\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643332 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643326 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-credential-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643352 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-workload-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643484 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643379 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/9017a28a-4013-4114-b8fb-4d3612657389-istio-token\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643632 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643490 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-istio-data\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643713 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643690 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-workload-certs\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643779 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643713 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-credential-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.643939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.643914 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/9017a28a-4013-4114-b8fb-4d3612657389-istiod-ca-cert\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.645417 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.645391 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/9017a28a-4013-4114-b8fb-4d3612657389-istio-envoy\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.645855 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.645830 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/9017a28a-4013-4114-b8fb-4d3612657389-istio-podinfo\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.651460 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.651440 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/9017a28a-4013-4114-b8fb-4d3612657389-istio-token\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.651641 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.651624 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrx6m\" (UniqueName: \"kubernetes.io/projected/9017a28a-4013-4114-b8fb-4d3612657389-kube-api-access-qrx6m\") pod \"maas-default-gateway-openshift-default-58b6f876-vbvzt\" (UID: \"9017a28a-4013-4114-b8fb-4d3612657389\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.827462 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.827370 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:30.955521 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.955497 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt"] Apr 21 14:39:30.957857 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:39:30.957830 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9017a28a_4013_4114_b8fb_4d3612657389.slice/crio-8ebac7f94bcc362a21680b3da1e569edd74f2762995a7871a3a5e3c1a6b68eb4 WatchSource:0}: Error finding container 8ebac7f94bcc362a21680b3da1e569edd74f2762995a7871a3a5e3c1a6b68eb4: Status 404 returned error can't find the container with id 8ebac7f94bcc362a21680b3da1e569edd74f2762995a7871a3a5e3c1a6b68eb4 Apr 21 14:39:30.960398 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.960367 2580 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892152Ki","pods":"250"} Apr 21 14:39:30.960477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.960431 2580 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892152Ki","pods":"250"} Apr 21 14:39:30.960477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:30.960461 2580 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892152Ki","pods":"250"} Apr 21 14:39:31.819989 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:31.819945 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" event={"ID":"9017a28a-4013-4114-b8fb-4d3612657389","Type":"ContainerStarted","Data":"85df0f55add6794faf7cb7ee6a2da895cd2aab40898a7e7b4cfd5bdeceb0db00"} Apr 21 14:39:31.819989 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:31.819991 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" event={"ID":"9017a28a-4013-4114-b8fb-4d3612657389","Type":"ContainerStarted","Data":"8ebac7f94bcc362a21680b3da1e569edd74f2762995a7871a3a5e3c1a6b68eb4"} Apr 21 14:39:31.828217 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:31.828192 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:31.842310 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:31.842262 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" podStartSLOduration=1.842241675 podStartE2EDuration="1.842241675s" podCreationTimestamp="2026-04-21 14:39:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:39:31.841212957 +0000 UTC m=+805.298697463" watchObservedRunningTime="2026-04-21 14:39:31.842241675 +0000 UTC m=+805.299726183" Apr 21 14:39:32.832823 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:32.832794 2580 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:33.826646 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:33.826619 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:33.827850 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:33.827830 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-vbvzt" Apr 21 14:39:35.014701 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.014668 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:39:35.018346 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.018323 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.020848 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.020826 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"default-dockercfg-h9qp9\"" Apr 21 14:39:35.020957 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.020867 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"limitador-limits-config-limitador\"" Apr 21 14:39:35.027552 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.027524 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:39:35.057316 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.057286 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:39:35.082687 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.082653 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kwjv\" (UniqueName: \"kubernetes.io/projected/8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d-kube-api-access-4kwjv\") pod \"limitador-limitador-78c99df468-xls6n\" (UID: \"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d\") " pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.082815 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.082740 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d-config-file\") pod \"limitador-limitador-78c99df468-xls6n\" (UID: \"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d\") " pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.184070 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.184031 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d-config-file\") pod \"limitador-limitador-78c99df468-xls6n\" (UID: \"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d\") " pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.184242 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.184103 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4kwjv\" (UniqueName: \"kubernetes.io/projected/8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d-kube-api-access-4kwjv\") pod \"limitador-limitador-78c99df468-xls6n\" (UID: \"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d\") " pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.184800 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.184778 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d-config-file\") pod \"limitador-limitador-78c99df468-xls6n\" (UID: \"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d\") " pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.191982 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.191950 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kwjv\" (UniqueName: \"kubernetes.io/projected/8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d-kube-api-access-4kwjv\") pod \"limitador-limitador-78c99df468-xls6n\" (UID: \"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d\") " pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.330258 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.330176 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:35.458642 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.458527 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:39:35.461645 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:39:35.461617 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b20ed5b_ca5f_41d5_aad2_1645e8f2d07d.slice/crio-821b622f88f5245fdf3614e01a6b5a362c335330eddb42b4e38191cd6d7e2f04 WatchSource:0}: Error finding container 821b622f88f5245fdf3614e01a6b5a362c335330eddb42b4e38191cd6d7e2f04: Status 404 returned error can't find the container with id 821b622f88f5245fdf3614e01a6b5a362c335330eddb42b4e38191cd6d7e2f04 Apr 21 14:39:35.485777 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.485747 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-tpbd4"] Apr 21 14:39:35.490680 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.490651 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" Apr 21 14:39:35.492892 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.492873 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-authorino-dockercfg-h5qxv\"" Apr 21 14:39:35.502275 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.502251 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-tpbd4"] Apr 21 14:39:35.588308 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.588209 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngzpp\" (UniqueName: \"kubernetes.io/projected/d90a9e4d-1cea-4022-8aea-d618c6c98a40-kube-api-access-ngzpp\") pod \"authorino-f99f4b5cd-tpbd4\" (UID: \"d90a9e4d-1cea-4022-8aea-d618c6c98a40\") " pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" Apr 21 14:39:35.689706 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.689661 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ngzpp\" (UniqueName: \"kubernetes.io/projected/d90a9e4d-1cea-4022-8aea-d618c6c98a40-kube-api-access-ngzpp\") pod \"authorino-f99f4b5cd-tpbd4\" (UID: \"d90a9e4d-1cea-4022-8aea-d618c6c98a40\") " pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" Apr 21 14:39:35.698849 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.698819 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngzpp\" (UniqueName: \"kubernetes.io/projected/d90a9e4d-1cea-4022-8aea-d618c6c98a40-kube-api-access-ngzpp\") pod \"authorino-f99f4b5cd-tpbd4\" (UID: \"d90a9e4d-1cea-4022-8aea-d618c6c98a40\") " pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" Apr 21 14:39:35.800024 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.799987 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" Apr 21 14:39:35.833751 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.833710 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" event={"ID":"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d","Type":"ContainerStarted","Data":"821b622f88f5245fdf3614e01a6b5a362c335330eddb42b4e38191cd6d7e2f04"} Apr 21 14:39:35.940875 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:35.940722 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-tpbd4"] Apr 21 14:39:35.944230 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:39:35.944184 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd90a9e4d_1cea_4022_8aea_d618c6c98a40.slice/crio-6a23af496a91ff378124b31b535f207318635615199395a5d45adf9a68ced471 WatchSource:0}: Error finding container 6a23af496a91ff378124b31b535f207318635615199395a5d45adf9a68ced471: Status 404 returned error can't find the container with id 6a23af496a91ff378124b31b535f207318635615199395a5d45adf9a68ced471 Apr 21 14:39:36.840828 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:36.840751 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" event={"ID":"d90a9e4d-1cea-4022-8aea-d618c6c98a40","Type":"ContainerStarted","Data":"6a23af496a91ff378124b31b535f207318635615199395a5d45adf9a68ced471"} Apr 21 14:39:40.857275 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:40.857235 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" event={"ID":"8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d","Type":"ContainerStarted","Data":"8102fd7c3c81cc9a2806605352c6fd21003d0cbe09ce27c689ba9187978951c8"} Apr 21 14:39:40.857829 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:40.857303 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:39:40.858560 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:40.858532 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" event={"ID":"d90a9e4d-1cea-4022-8aea-d618c6c98a40","Type":"ContainerStarted","Data":"8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa"} Apr 21 14:39:40.881244 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:40.881197 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" podStartSLOduration=2.465035859 podStartE2EDuration="6.881183753s" podCreationTimestamp="2026-04-21 14:39:34 +0000 UTC" firstStartedPulling="2026-04-21 14:39:35.463436882 +0000 UTC m=+808.920921367" lastFinishedPulling="2026-04-21 14:39:39.879584758 +0000 UTC m=+813.337069261" observedRunningTime="2026-04-21 14:39:40.878350984 +0000 UTC m=+814.335835488" watchObservedRunningTime="2026-04-21 14:39:40.881183753 +0000 UTC m=+814.338668259" Apr 21 14:39:40.897115 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:40.897070 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" podStartSLOduration=1.963281838 podStartE2EDuration="5.897057545s" podCreationTimestamp="2026-04-21 14:39:35 +0000 UTC" firstStartedPulling="2026-04-21 14:39:35.946149406 +0000 UTC m=+809.403633908" lastFinishedPulling="2026-04-21 14:39:39.879925121 +0000 UTC m=+813.337409615" observedRunningTime="2026-04-21 14:39:40.895055194 +0000 UTC m=+814.352539700" watchObservedRunningTime="2026-04-21 14:39:40.897057545 +0000 UTC m=+814.354542051" Apr 21 14:39:41.720486 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:41.720447 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-tpbd4"] Apr 21 14:39:42.865784 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:42.865742 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" podUID="d90a9e4d-1cea-4022-8aea-d618c6c98a40" containerName="authorino" containerID="cri-o://8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa" gracePeriod=30 Apr 21 14:39:43.110820 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.110795 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" Apr 21 14:39:43.160178 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.160098 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngzpp\" (UniqueName: \"kubernetes.io/projected/d90a9e4d-1cea-4022-8aea-d618c6c98a40-kube-api-access-ngzpp\") pod \"d90a9e4d-1cea-4022-8aea-d618c6c98a40\" (UID: \"d90a9e4d-1cea-4022-8aea-d618c6c98a40\") " Apr 21 14:39:43.162304 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.162280 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d90a9e4d-1cea-4022-8aea-d618c6c98a40-kube-api-access-ngzpp" (OuterVolumeSpecName: "kube-api-access-ngzpp") pod "d90a9e4d-1cea-4022-8aea-d618c6c98a40" (UID: "d90a9e4d-1cea-4022-8aea-d618c6c98a40"). InnerVolumeSpecName "kube-api-access-ngzpp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:39:43.260649 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.260622 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ngzpp\" (UniqueName: \"kubernetes.io/projected/d90a9e4d-1cea-4022-8aea-d618c6c98a40-kube-api-access-ngzpp\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:39:43.870263 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.870221 2580 generic.go:358] "Generic (PLEG): container finished" podID="d90a9e4d-1cea-4022-8aea-d618c6c98a40" containerID="8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa" exitCode=0 Apr 21 14:39:43.870754 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.870276 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" Apr 21 14:39:43.870754 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.870316 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" event={"ID":"d90a9e4d-1cea-4022-8aea-d618c6c98a40","Type":"ContainerDied","Data":"8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa"} Apr 21 14:39:43.870754 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.870353 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-tpbd4" event={"ID":"d90a9e4d-1cea-4022-8aea-d618c6c98a40","Type":"ContainerDied","Data":"6a23af496a91ff378124b31b535f207318635615199395a5d45adf9a68ced471"} Apr 21 14:39:43.870754 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.870374 2580 scope.go:117] "RemoveContainer" containerID="8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa" Apr 21 14:39:43.878451 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.878431 2580 scope.go:117] "RemoveContainer" containerID="8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa" Apr 21 14:39:43.878724 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:39:43.878705 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa\": container with ID starting with 8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa not found: ID does not exist" containerID="8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa" Apr 21 14:39:43.878788 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.878736 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa"} err="failed to get container status \"8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa\": rpc error: code = NotFound desc = could not find container \"8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa\": container with ID starting with 8e323ee3c63ba84b0ab717bec1f3d391d7f6d285e9da69854abf2c47c42594aa not found: ID does not exist" Apr 21 14:39:43.887142 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.887119 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-tpbd4"] Apr 21 14:39:43.888870 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:43.888848 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-tpbd4"] Apr 21 14:39:45.246120 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:45.246085 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d90a9e4d-1cea-4022-8aea-d618c6c98a40" path="/var/lib/kubelet/pods/d90a9e4d-1cea-4022-8aea-d618c6c98a40/volumes" Apr 21 14:39:51.863310 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:39:51.863274 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/limitador-limitador-78c99df468-xls6n" Apr 21 14:41:05.443644 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.443605 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-8b475cf9f-hhd4c"] Apr 21 14:41:05.444069 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.443928 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d90a9e4d-1cea-4022-8aea-d618c6c98a40" containerName="authorino" Apr 21 14:41:05.444069 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.443939 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="d90a9e4d-1cea-4022-8aea-d618c6c98a40" containerName="authorino" Apr 21 14:41:05.444069 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.444006 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="d90a9e4d-1cea-4022-8aea-d618c6c98a40" containerName="authorino" Apr 21 14:41:05.446980 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.446962 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-hhd4c" Apr 21 14:41:05.449366 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.449337 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-authorino-dockercfg-h5qxv\"" Apr 21 14:41:05.452764 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.452737 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-hhd4c"] Apr 21 14:41:05.575679 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.575632 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lm5k\" (UniqueName: \"kubernetes.io/projected/3dcbd661-b8b7-4b5b-a5d1-274607ea23e5-kube-api-access-4lm5k\") pod \"authorino-8b475cf9f-hhd4c\" (UID: \"3dcbd661-b8b7-4b5b-a5d1-274607ea23e5\") " pod="kuadrant-system/authorino-8b475cf9f-hhd4c" Apr 21 14:41:05.670201 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.670160 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-hhd4c"] Apr 21 14:41:05.670492 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:41:05.670467 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-4lm5k], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kuadrant-system/authorino-8b475cf9f-hhd4c" podUID="3dcbd661-b8b7-4b5b-a5d1-274607ea23e5" Apr 21 14:41:05.676774 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.676738 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4lm5k\" (UniqueName: \"kubernetes.io/projected/3dcbd661-b8b7-4b5b-a5d1-274607ea23e5-kube-api-access-4lm5k\") pod \"authorino-8b475cf9f-hhd4c\" (UID: \"3dcbd661-b8b7-4b5b-a5d1-274607ea23e5\") " pod="kuadrant-system/authorino-8b475cf9f-hhd4c" Apr 21 14:41:05.691099 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.691068 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lm5k\" (UniqueName: \"kubernetes.io/projected/3dcbd661-b8b7-4b5b-a5d1-274607ea23e5-kube-api-access-4lm5k\") pod \"authorino-8b475cf9f-hhd4c\" (UID: \"3dcbd661-b8b7-4b5b-a5d1-274607ea23e5\") " pod="kuadrant-system/authorino-8b475cf9f-hhd4c" Apr 21 14:41:05.695723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.695654 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-57547b964d-mhqq8"] Apr 21 14:41:05.699136 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.699115 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-57547b964d-mhqq8" Apr 21 14:41:05.707520 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.707493 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-57547b964d-mhqq8"] Apr 21 14:41:05.878477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.878434 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9sk5\" (UniqueName: \"kubernetes.io/projected/3b100418-ad12-4a21-a5bb-c4aa589c57c5-kube-api-access-p9sk5\") pod \"authorino-57547b964d-mhqq8\" (UID: \"3b100418-ad12-4a21-a5bb-c4aa589c57c5\") " pod="kuadrant-system/authorino-57547b964d-mhqq8" Apr 21 14:41:05.979126 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.979036 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p9sk5\" (UniqueName: \"kubernetes.io/projected/3b100418-ad12-4a21-a5bb-c4aa589c57c5-kube-api-access-p9sk5\") pod \"authorino-57547b964d-mhqq8\" (UID: \"3b100418-ad12-4a21-a5bb-c4aa589c57c5\") " pod="kuadrant-system/authorino-57547b964d-mhqq8" Apr 21 14:41:05.988258 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:05.988234 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9sk5\" (UniqueName: \"kubernetes.io/projected/3b100418-ad12-4a21-a5bb-c4aa589c57c5-kube-api-access-p9sk5\") pod \"authorino-57547b964d-mhqq8\" (UID: \"3b100418-ad12-4a21-a5bb-c4aa589c57c5\") " pod="kuadrant-system/authorino-57547b964d-mhqq8" Apr 21 14:41:06.009896 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.009860 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-57547b964d-mhqq8" Apr 21 14:41:06.134723 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.134694 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-57547b964d-mhqq8"] Apr 21 14:41:06.145483 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.145456 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-57547b964d-mhqq8" event={"ID":"3b100418-ad12-4a21-a5bb-c4aa589c57c5","Type":"ContainerStarted","Data":"94c3e13f6f28ccc56701512217667183bcea172dd199603b7b43f2a1eb44df6b"} Apr 21 14:41:06.145611 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.145499 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-hhd4c" Apr 21 14:41:06.150545 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.150518 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-hhd4c" Apr 21 14:41:06.281978 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.281892 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lm5k\" (UniqueName: \"kubernetes.io/projected/3dcbd661-b8b7-4b5b-a5d1-274607ea23e5-kube-api-access-4lm5k\") pod \"3dcbd661-b8b7-4b5b-a5d1-274607ea23e5\" (UID: \"3dcbd661-b8b7-4b5b-a5d1-274607ea23e5\") " Apr 21 14:41:06.284218 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.284183 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dcbd661-b8b7-4b5b-a5d1-274607ea23e5-kube-api-access-4lm5k" (OuterVolumeSpecName: "kube-api-access-4lm5k") pod "3dcbd661-b8b7-4b5b-a5d1-274607ea23e5" (UID: "3dcbd661-b8b7-4b5b-a5d1-274607ea23e5"). InnerVolumeSpecName "kube-api-access-4lm5k". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:41:06.382863 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.382831 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4lm5k\" (UniqueName: \"kubernetes.io/projected/3dcbd661-b8b7-4b5b-a5d1-274607ea23e5-kube-api-access-4lm5k\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:41:06.461495 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.461462 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-57547b964d-mhqq8"] Apr 21 14:41:06.493287 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.493255 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-b865f9f7b-ls2zr"] Apr 21 14:41:06.497733 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.497715 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.500053 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.500028 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-server-cert\"" Apr 21 14:41:06.507067 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.507043 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-b865f9f7b-ls2zr"] Apr 21 14:41:06.686282 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.686246 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2x9b\" (UniqueName: \"kubernetes.io/projected/97a60321-437f-484d-a535-ca02c03d8b11-kube-api-access-d2x9b\") pod \"authorino-b865f9f7b-ls2zr\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.686447 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.686306 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/97a60321-437f-484d-a535-ca02c03d8b11-tls-cert\") pod \"authorino-b865f9f7b-ls2zr\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.787042 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.787011 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d2x9b\" (UniqueName: \"kubernetes.io/projected/97a60321-437f-484d-a535-ca02c03d8b11-kube-api-access-d2x9b\") pod \"authorino-b865f9f7b-ls2zr\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.787224 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.787093 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/97a60321-437f-484d-a535-ca02c03d8b11-tls-cert\") pod \"authorino-b865f9f7b-ls2zr\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.789993 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.789966 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/97a60321-437f-484d-a535-ca02c03d8b11-tls-cert\") pod \"authorino-b865f9f7b-ls2zr\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.796156 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.796128 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2x9b\" (UniqueName: \"kubernetes.io/projected/97a60321-437f-484d-a535-ca02c03d8b11-kube-api-access-d2x9b\") pod \"authorino-b865f9f7b-ls2zr\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.806851 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.806816 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:41:06.947526 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:06.947448 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-b865f9f7b-ls2zr"] Apr 21 14:41:06.950780 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:41:06.950750 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97a60321_437f_484d_a535_ca02c03d8b11.slice/crio-1a346d5072a4a7924b8b23c14a02251518c0dfb30349a8a77bc93371d2a91abe WatchSource:0}: Error finding container 1a346d5072a4a7924b8b23c14a02251518c0dfb30349a8a77bc93371d2a91abe: Status 404 returned error can't find the container with id 1a346d5072a4a7924b8b23c14a02251518c0dfb30349a8a77bc93371d2a91abe Apr 21 14:41:07.149354 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.149315 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-57547b964d-mhqq8" event={"ID":"3b100418-ad12-4a21-a5bb-c4aa589c57c5","Type":"ContainerStarted","Data":"5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96"} Apr 21 14:41:07.149548 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.149390 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-57547b964d-mhqq8" podUID="3b100418-ad12-4a21-a5bb-c4aa589c57c5" containerName="authorino" containerID="cri-o://5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96" gracePeriod=30 Apr 21 14:41:07.150597 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.150547 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" event={"ID":"97a60321-437f-484d-a535-ca02c03d8b11","Type":"ContainerStarted","Data":"1a346d5072a4a7924b8b23c14a02251518c0dfb30349a8a77bc93371d2a91abe"} Apr 21 14:41:07.150707 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.150596 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-hhd4c" Apr 21 14:41:07.155794 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.155772 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:41:07.157493 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.157474 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:41:07.157696 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.157678 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:41:07.159338 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.159302 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:41:07.161236 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.161219 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:41:07.162662 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.162645 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:41:07.165471 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.165426 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-57547b964d-mhqq8" podStartSLOduration=1.593547385 podStartE2EDuration="2.165410742s" podCreationTimestamp="2026-04-21 14:41:05 +0000 UTC" firstStartedPulling="2026-04-21 14:41:06.13981386 +0000 UTC m=+899.597298346" lastFinishedPulling="2026-04-21 14:41:06.711677204 +0000 UTC m=+900.169161703" observedRunningTime="2026-04-21 14:41:07.16394244 +0000 UTC m=+900.621426947" watchObservedRunningTime="2026-04-21 14:41:07.165410742 +0000 UTC m=+900.622895252" Apr 21 14:41:07.185610 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.185576 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-hhd4c"] Apr 21 14:41:07.190555 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.190525 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-hhd4c"] Apr 21 14:41:07.247169 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.247096 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dcbd661-b8b7-4b5b-a5d1-274607ea23e5" path="/var/lib/kubelet/pods/3dcbd661-b8b7-4b5b-a5d1-274607ea23e5/volumes" Apr 21 14:41:07.402750 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.402726 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-57547b964d-mhqq8" Apr 21 14:41:07.492129 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.492097 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9sk5\" (UniqueName: \"kubernetes.io/projected/3b100418-ad12-4a21-a5bb-c4aa589c57c5-kube-api-access-p9sk5\") pod \"3b100418-ad12-4a21-a5bb-c4aa589c57c5\" (UID: \"3b100418-ad12-4a21-a5bb-c4aa589c57c5\") " Apr 21 14:41:07.494281 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.494256 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b100418-ad12-4a21-a5bb-c4aa589c57c5-kube-api-access-p9sk5" (OuterVolumeSpecName: "kube-api-access-p9sk5") pod "3b100418-ad12-4a21-a5bb-c4aa589c57c5" (UID: "3b100418-ad12-4a21-a5bb-c4aa589c57c5"). InnerVolumeSpecName "kube-api-access-p9sk5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:41:07.593701 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:07.593612 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-p9sk5\" (UniqueName: \"kubernetes.io/projected/3b100418-ad12-4a21-a5bb-c4aa589c57c5-kube-api-access-p9sk5\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:41:08.155509 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.155467 2580 generic.go:358] "Generic (PLEG): container finished" podID="3b100418-ad12-4a21-a5bb-c4aa589c57c5" containerID="5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96" exitCode=0 Apr 21 14:41:08.155690 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.155521 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-57547b964d-mhqq8" Apr 21 14:41:08.155690 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.155550 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-57547b964d-mhqq8" event={"ID":"3b100418-ad12-4a21-a5bb-c4aa589c57c5","Type":"ContainerDied","Data":"5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96"} Apr 21 14:41:08.155690 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.155615 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-57547b964d-mhqq8" event={"ID":"3b100418-ad12-4a21-a5bb-c4aa589c57c5","Type":"ContainerDied","Data":"94c3e13f6f28ccc56701512217667183bcea172dd199603b7b43f2a1eb44df6b"} Apr 21 14:41:08.155690 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.155636 2580 scope.go:117] "RemoveContainer" containerID="5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96" Apr 21 14:41:08.156988 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.156965 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" event={"ID":"97a60321-437f-484d-a535-ca02c03d8b11","Type":"ContainerStarted","Data":"ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24"} Apr 21 14:41:08.164355 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.164336 2580 scope.go:117] "RemoveContainer" containerID="5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96" Apr 21 14:41:08.164704 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:41:08.164686 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96\": container with ID starting with 5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96 not found: ID does not exist" containerID="5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96" Apr 21 14:41:08.164744 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.164714 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96"} err="failed to get container status \"5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96\": rpc error: code = NotFound desc = could not find container \"5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96\": container with ID starting with 5dee0fe75065c6f29f0b36a96ab82558aa70f4c67f874506016ec5b538411d96 not found: ID does not exist" Apr 21 14:41:08.175661 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.175619 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" podStartSLOduration=1.836432589 podStartE2EDuration="2.175607957s" podCreationTimestamp="2026-04-21 14:41:06 +0000 UTC" firstStartedPulling="2026-04-21 14:41:06.952157705 +0000 UTC m=+900.409642190" lastFinishedPulling="2026-04-21 14:41:07.291333071 +0000 UTC m=+900.748817558" observedRunningTime="2026-04-21 14:41:08.174228929 +0000 UTC m=+901.631713436" watchObservedRunningTime="2026-04-21 14:41:08.175607957 +0000 UTC m=+901.633092464" Apr 21 14:41:08.192427 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.192400 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-57547b964d-mhqq8"] Apr 21 14:41:08.198316 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:08.198286 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-57547b964d-mhqq8"] Apr 21 14:41:09.085829 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.085795 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-ljrhs"] Apr 21 14:41:09.086211 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.086152 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3b100418-ad12-4a21-a5bb-c4aa589c57c5" containerName="authorino" Apr 21 14:41:09.086211 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.086162 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b100418-ad12-4a21-a5bb-c4aa589c57c5" containerName="authorino" Apr 21 14:41:09.086281 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.086223 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="3b100418-ad12-4a21-a5bb-c4aa589c57c5" containerName="authorino" Apr 21 14:41:09.090557 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.090534 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:09.094187 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.094168 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-controller-dockercfg-mqm8t\"" Apr 21 14:41:09.102810 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.102785 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-ljrhs"] Apr 21 14:41:09.106279 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.106256 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxd8z\" (UniqueName: \"kubernetes.io/projected/0b930fa3-1c7c-4a78-972e-f42e13ecd943-kube-api-access-mxd8z\") pod \"maas-controller-6d4c8f55f9-ljrhs\" (UID: \"0b930fa3-1c7c-4a78-972e-f42e13ecd943\") " pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:09.207007 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.206958 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mxd8z\" (UniqueName: \"kubernetes.io/projected/0b930fa3-1c7c-4a78-972e-f42e13ecd943-kube-api-access-mxd8z\") pod \"maas-controller-6d4c8f55f9-ljrhs\" (UID: \"0b930fa3-1c7c-4a78-972e-f42e13ecd943\") " pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:09.217883 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.217856 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxd8z\" (UniqueName: \"kubernetes.io/projected/0b930fa3-1c7c-4a78-972e-f42e13ecd943-kube-api-access-mxd8z\") pod \"maas-controller-6d4c8f55f9-ljrhs\" (UID: \"0b930fa3-1c7c-4a78-972e-f42e13ecd943\") " pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:09.246975 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.246939 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b100418-ad12-4a21-a5bb-c4aa589c57c5" path="/var/lib/kubelet/pods/3b100418-ad12-4a21-a5bb-c4aa589c57c5/volumes" Apr 21 14:41:09.261715 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.261680 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-666f94d88c-m72jd"] Apr 21 14:41:09.265191 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.265173 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-666f94d88c-m72jd" Apr 21 14:41:09.277146 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.277122 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-666f94d88c-m72jd"] Apr 21 14:41:09.307965 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.307917 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5z7w\" (UniqueName: \"kubernetes.io/projected/9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd-kube-api-access-r5z7w\") pod \"maas-controller-666f94d88c-m72jd\" (UID: \"9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd\") " pod="opendatahub/maas-controller-666f94d88c-m72jd" Apr 21 14:41:09.398053 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.397967 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-666f94d88c-m72jd"] Apr 21 14:41:09.398237 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:41:09.398216 2580 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-r5z7w], unattached volumes=[], failed to process volumes=[]: context canceled" pod="opendatahub/maas-controller-666f94d88c-m72jd" podUID="9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd" Apr 21 14:41:09.401308 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.401288 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:09.409409 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.409367 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r5z7w\" (UniqueName: \"kubernetes.io/projected/9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd-kube-api-access-r5z7w\") pod \"maas-controller-666f94d88c-m72jd\" (UID: \"9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd\") " pod="opendatahub/maas-controller-666f94d88c-m72jd" Apr 21 14:41:09.423438 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.423413 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5z7w\" (UniqueName: \"kubernetes.io/projected/9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd-kube-api-access-r5z7w\") pod \"maas-controller-666f94d88c-m72jd\" (UID: \"9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd\") " pod="opendatahub/maas-controller-666f94d88c-m72jd" Apr 21 14:41:09.428035 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.428005 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-6d5bc4f566-tpg7r"] Apr 21 14:41:09.432486 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.432463 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:09.440941 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.440914 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6d5bc4f566-tpg7r"] Apr 21 14:41:09.510869 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.510835 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdj5s\" (UniqueName: \"kubernetes.io/projected/d2c7f535-1a26-4398-a262-a93365e992f3-kube-api-access-tdj5s\") pod \"maas-controller-6d5bc4f566-tpg7r\" (UID: \"d2c7f535-1a26-4398-a262-a93365e992f3\") " pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:09.536546 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.536521 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-ljrhs"] Apr 21 14:41:09.538605 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:41:09.538559 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b930fa3_1c7c_4a78_972e_f42e13ecd943.slice/crio-b430d0a35855957d1b347488828e1eaf9943c91217d99d3d5c348cd0d0821b38 WatchSource:0}: Error finding container b430d0a35855957d1b347488828e1eaf9943c91217d99d3d5c348cd0d0821b38: Status 404 returned error can't find the container with id b430d0a35855957d1b347488828e1eaf9943c91217d99d3d5c348cd0d0821b38 Apr 21 14:41:09.611952 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.611918 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tdj5s\" (UniqueName: \"kubernetes.io/projected/d2c7f535-1a26-4398-a262-a93365e992f3-kube-api-access-tdj5s\") pod \"maas-controller-6d5bc4f566-tpg7r\" (UID: \"d2c7f535-1a26-4398-a262-a93365e992f3\") " pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:09.620155 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.620134 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdj5s\" (UniqueName: \"kubernetes.io/projected/d2c7f535-1a26-4398-a262-a93365e992f3-kube-api-access-tdj5s\") pod \"maas-controller-6d5bc4f566-tpg7r\" (UID: \"d2c7f535-1a26-4398-a262-a93365e992f3\") " pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:09.745860 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.745825 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:09.873337 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:09.873296 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6d5bc4f566-tpg7r"] Apr 21 14:41:09.875478 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:41:09.875448 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2c7f535_1a26_4398_a262_a93365e992f3.slice/crio-6dece40ebedba1ccc4756603e044080fd99dd716b06324b697abc7b41a6b149b WatchSource:0}: Error finding container 6dece40ebedba1ccc4756603e044080fd99dd716b06324b697abc7b41a6b149b: Status 404 returned error can't find the container with id 6dece40ebedba1ccc4756603e044080fd99dd716b06324b697abc7b41a6b149b Apr 21 14:41:10.167636 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:10.167595 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" event={"ID":"0b930fa3-1c7c-4a78-972e-f42e13ecd943","Type":"ContainerStarted","Data":"b430d0a35855957d1b347488828e1eaf9943c91217d99d3d5c348cd0d0821b38"} Apr 21 14:41:10.169335 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:10.169312 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-666f94d88c-m72jd" Apr 21 14:41:10.169751 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:10.169728 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" event={"ID":"d2c7f535-1a26-4398-a262-a93365e992f3","Type":"ContainerStarted","Data":"6dece40ebedba1ccc4756603e044080fd99dd716b06324b697abc7b41a6b149b"} Apr 21 14:41:10.176207 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:10.176184 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-666f94d88c-m72jd" Apr 21 14:41:10.217041 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:10.217006 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5z7w\" (UniqueName: \"kubernetes.io/projected/9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd-kube-api-access-r5z7w\") pod \"9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd\" (UID: \"9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd\") " Apr 21 14:41:10.220240 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:10.220205 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd-kube-api-access-r5z7w" (OuterVolumeSpecName: "kube-api-access-r5z7w") pod "9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd" (UID: "9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd"). InnerVolumeSpecName "kube-api-access-r5z7w". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:41:10.318064 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:10.318024 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-r5z7w\" (UniqueName: \"kubernetes.io/projected/9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd-kube-api-access-r5z7w\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:41:11.174415 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:11.174381 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-666f94d88c-m72jd" Apr 21 14:41:11.218257 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:11.218225 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-666f94d88c-m72jd"] Apr 21 14:41:11.221905 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:11.221869 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-666f94d88c-m72jd"] Apr 21 14:41:11.248108 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:11.248057 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd" path="/var/lib/kubelet/pods/9d2ac1e9-2a9a-4f94-a0cd-68e3094f6ccd/volumes" Apr 21 14:41:13.185101 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:13.185058 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" event={"ID":"d2c7f535-1a26-4398-a262-a93365e992f3","Type":"ContainerStarted","Data":"1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b"} Apr 21 14:41:13.185586 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:13.185118 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:13.186528 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:13.186504 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" event={"ID":"0b930fa3-1c7c-4a78-972e-f42e13ecd943","Type":"ContainerStarted","Data":"a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e"} Apr 21 14:41:13.186668 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:13.186653 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:13.211407 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:13.211354 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" podStartSLOduration=1.595640463 podStartE2EDuration="4.211341318s" podCreationTimestamp="2026-04-21 14:41:09 +0000 UTC" firstStartedPulling="2026-04-21 14:41:09.876755179 +0000 UTC m=+903.334239664" lastFinishedPulling="2026-04-21 14:41:12.49245602 +0000 UTC m=+905.949940519" observedRunningTime="2026-04-21 14:41:13.209248261 +0000 UTC m=+906.666732796" watchObservedRunningTime="2026-04-21 14:41:13.211341318 +0000 UTC m=+906.668825891" Apr 21 14:41:13.232226 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:13.232175 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" podStartSLOduration=1.283391386 podStartE2EDuration="4.232160147s" podCreationTimestamp="2026-04-21 14:41:09 +0000 UTC" firstStartedPulling="2026-04-21 14:41:09.540037203 +0000 UTC m=+902.997521688" lastFinishedPulling="2026-04-21 14:41:12.488805961 +0000 UTC m=+905.946290449" observedRunningTime="2026-04-21 14:41:13.230748382 +0000 UTC m=+906.688232889" watchObservedRunningTime="2026-04-21 14:41:13.232160147 +0000 UTC m=+906.689645032" Apr 21 14:41:15.319169 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:15.319129 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:41:24.198336 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.198302 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:24.198868 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.198463 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:24.281372 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.281327 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-ljrhs"] Apr 21 14:41:24.281676 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.281630 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" podUID="0b930fa3-1c7c-4a78-972e-f42e13ecd943" containerName="manager" containerID="cri-o://a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e" gracePeriod=10 Apr 21 14:41:24.529992 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.529971 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:24.581453 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.581413 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-744d89cb86-bkwjx"] Apr 21 14:41:24.582033 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.582012 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0b930fa3-1c7c-4a78-972e-f42e13ecd943" containerName="manager" Apr 21 14:41:24.582091 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.582040 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b930fa3-1c7c-4a78-972e-f42e13ecd943" containerName="manager" Apr 21 14:41:24.582182 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.582169 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="0b930fa3-1c7c-4a78-972e-f42e13ecd943" containerName="manager" Apr 21 14:41:24.585460 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.585444 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:41:24.593322 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.593298 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-744d89cb86-bkwjx"] Apr 21 14:41:24.648417 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.648374 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxd8z\" (UniqueName: \"kubernetes.io/projected/0b930fa3-1c7c-4a78-972e-f42e13ecd943-kube-api-access-mxd8z\") pod \"0b930fa3-1c7c-4a78-972e-f42e13ecd943\" (UID: \"0b930fa3-1c7c-4a78-972e-f42e13ecd943\") " Apr 21 14:41:24.648614 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.648549 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfh7t\" (UniqueName: \"kubernetes.io/projected/7adf31a4-3dfc-451d-b65d-66d6d8ff93d7-kube-api-access-nfh7t\") pod \"maas-controller-744d89cb86-bkwjx\" (UID: \"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7\") " pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:41:24.650747 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.650722 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b930fa3-1c7c-4a78-972e-f42e13ecd943-kube-api-access-mxd8z" (OuterVolumeSpecName: "kube-api-access-mxd8z") pod "0b930fa3-1c7c-4a78-972e-f42e13ecd943" (UID: "0b930fa3-1c7c-4a78-972e-f42e13ecd943"). InnerVolumeSpecName "kube-api-access-mxd8z". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:41:24.749766 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.749724 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nfh7t\" (UniqueName: \"kubernetes.io/projected/7adf31a4-3dfc-451d-b65d-66d6d8ff93d7-kube-api-access-nfh7t\") pod \"maas-controller-744d89cb86-bkwjx\" (UID: \"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7\") " pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:41:24.749939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.749779 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mxd8z\" (UniqueName: \"kubernetes.io/projected/0b930fa3-1c7c-4a78-972e-f42e13ecd943-kube-api-access-mxd8z\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:41:24.758513 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.758484 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfh7t\" (UniqueName: \"kubernetes.io/projected/7adf31a4-3dfc-451d-b65d-66d6d8ff93d7-kube-api-access-nfh7t\") pod \"maas-controller-744d89cb86-bkwjx\" (UID: \"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7\") " pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:41:24.896689 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:24.896592 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:41:25.025046 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.025019 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-744d89cb86-bkwjx"] Apr 21 14:41:25.026725 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:41:25.026687 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7adf31a4_3dfc_451d_b65d_66d6d8ff93d7.slice/crio-b8d5ba90d2a70fc71de04ae8686cf132a55e7d9ddbcfaa87fcb521fd62bc1207 WatchSource:0}: Error finding container b8d5ba90d2a70fc71de04ae8686cf132a55e7d9ddbcfaa87fcb521fd62bc1207: Status 404 returned error can't find the container with id b8d5ba90d2a70fc71de04ae8686cf132a55e7d9ddbcfaa87fcb521fd62bc1207 Apr 21 14:41:25.228097 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.228062 2580 generic.go:358] "Generic (PLEG): container finished" podID="0b930fa3-1c7c-4a78-972e-f42e13ecd943" containerID="a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e" exitCode=0 Apr 21 14:41:25.228613 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.228143 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" Apr 21 14:41:25.228613 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.228153 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" event={"ID":"0b930fa3-1c7c-4a78-972e-f42e13ecd943","Type":"ContainerDied","Data":"a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e"} Apr 21 14:41:25.228613 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.228187 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-ljrhs" event={"ID":"0b930fa3-1c7c-4a78-972e-f42e13ecd943","Type":"ContainerDied","Data":"b430d0a35855957d1b347488828e1eaf9943c91217d99d3d5c348cd0d0821b38"} Apr 21 14:41:25.228613 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.228205 2580 scope.go:117] "RemoveContainer" containerID="a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e" Apr 21 14:41:25.229271 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.229224 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-744d89cb86-bkwjx" event={"ID":"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7","Type":"ContainerStarted","Data":"b8d5ba90d2a70fc71de04ae8686cf132a55e7d9ddbcfaa87fcb521fd62bc1207"} Apr 21 14:41:25.237134 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.237117 2580 scope.go:117] "RemoveContainer" containerID="a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e" Apr 21 14:41:25.237412 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:41:25.237394 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e\": container with ID starting with a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e not found: ID does not exist" containerID="a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e" Apr 21 14:41:25.237451 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.237422 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e"} err="failed to get container status \"a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e\": rpc error: code = NotFound desc = could not find container \"a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e\": container with ID starting with a8c9e116507b223cfdc6a30830cfbfd2cfe11977bde9853973e1429c03abcb6e not found: ID does not exist" Apr 21 14:41:25.249105 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.249080 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-ljrhs"] Apr 21 14:41:25.255833 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:25.255810 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-ljrhs"] Apr 21 14:41:26.234164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:26.234126 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-744d89cb86-bkwjx" event={"ID":"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7","Type":"ContainerStarted","Data":"be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c"} Apr 21 14:41:26.234557 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:26.234283 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:41:26.252100 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:26.252047 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-744d89cb86-bkwjx" podStartSLOduration=1.9394311979999999 podStartE2EDuration="2.252033275s" podCreationTimestamp="2026-04-21 14:41:24 +0000 UTC" firstStartedPulling="2026-04-21 14:41:25.028245013 +0000 UTC m=+918.485729500" lastFinishedPulling="2026-04-21 14:41:25.340847089 +0000 UTC m=+918.798331577" observedRunningTime="2026-04-21 14:41:26.250107873 +0000 UTC m=+919.707592380" watchObservedRunningTime="2026-04-21 14:41:26.252033275 +0000 UTC m=+919.709517782" Apr 21 14:41:27.245103 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:27.245072 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b930fa3-1c7c-4a78-972e-f42e13ecd943" path="/var/lib/kubelet/pods/0b930fa3-1c7c-4a78-972e-f42e13ecd943/volumes" Apr 21 14:41:37.246109 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:37.246075 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:41:37.295372 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:37.295338 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d5bc4f566-tpg7r"] Apr 21 14:41:37.295655 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:37.295629 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" podUID="d2c7f535-1a26-4398-a262-a93365e992f3" containerName="manager" containerID="cri-o://1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b" gracePeriod=10 Apr 21 14:41:37.541128 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:37.541106 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:37.662944 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:37.662901 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdj5s\" (UniqueName: \"kubernetes.io/projected/d2c7f535-1a26-4398-a262-a93365e992f3-kube-api-access-tdj5s\") pod \"d2c7f535-1a26-4398-a262-a93365e992f3\" (UID: \"d2c7f535-1a26-4398-a262-a93365e992f3\") " Apr 21 14:41:37.665141 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:37.665106 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2c7f535-1a26-4398-a262-a93365e992f3-kube-api-access-tdj5s" (OuterVolumeSpecName: "kube-api-access-tdj5s") pod "d2c7f535-1a26-4398-a262-a93365e992f3" (UID: "d2c7f535-1a26-4398-a262-a93365e992f3"). InnerVolumeSpecName "kube-api-access-tdj5s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:41:37.763977 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:37.763939 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tdj5s\" (UniqueName: \"kubernetes.io/projected/d2c7f535-1a26-4398-a262-a93365e992f3-kube-api-access-tdj5s\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:41:38.276833 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.276798 2580 generic.go:358] "Generic (PLEG): container finished" podID="d2c7f535-1a26-4398-a262-a93365e992f3" containerID="1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b" exitCode=0 Apr 21 14:41:38.277237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.276861 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" Apr 21 14:41:38.277237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.276886 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" event={"ID":"d2c7f535-1a26-4398-a262-a93365e992f3","Type":"ContainerDied","Data":"1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b"} Apr 21 14:41:38.277237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.276924 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d5bc4f566-tpg7r" event={"ID":"d2c7f535-1a26-4398-a262-a93365e992f3","Type":"ContainerDied","Data":"6dece40ebedba1ccc4756603e044080fd99dd716b06324b697abc7b41a6b149b"} Apr 21 14:41:38.277237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.276939 2580 scope.go:117] "RemoveContainer" containerID="1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b" Apr 21 14:41:38.285725 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.285705 2580 scope.go:117] "RemoveContainer" containerID="1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b" Apr 21 14:41:38.285989 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:41:38.285966 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b\": container with ID starting with 1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b not found: ID does not exist" containerID="1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b" Apr 21 14:41:38.286055 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.285997 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b"} err="failed to get container status \"1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b\": rpc error: code = NotFound desc = could not find container \"1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b\": container with ID starting with 1ed37e7496622cdac8a853161925191f5cf264df0e3bf7fd163efc130ff4168b not found: ID does not exist" Apr 21 14:41:38.297896 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.297863 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d5bc4f566-tpg7r"] Apr 21 14:41:38.303607 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:38.303584 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-6d5bc4f566-tpg7r"] Apr 21 14:41:39.247644 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:39.247601 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2c7f535-1a26-4398-a262-a93365e992f3" path="/var/lib/kubelet/pods/d2c7f535-1a26-4398-a262-a93365e992f3/volumes" Apr 21 14:41:42.739325 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.739285 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-api-5c64fc8cbc-w9f4j"] Apr 21 14:41:42.739769 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.739663 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d2c7f535-1a26-4398-a262-a93365e992f3" containerName="manager" Apr 21 14:41:42.739769 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.739676 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2c7f535-1a26-4398-a262-a93365e992f3" containerName="manager" Apr 21 14:41:42.739769 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.739732 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="d2c7f535-1a26-4398-a262-a93365e992f3" containerName="manager" Apr 21 14:41:42.743929 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.743908 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:42.746642 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.746623 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-api-dockercfg-h4b5p\"" Apr 21 14:41:42.746917 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.746901 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"maas-parameters\"" Apr 21 14:41:42.747293 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.747273 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-api-serving-cert\"" Apr 21 14:41:42.760413 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.760385 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-api-5c64fc8cbc-w9f4j"] Apr 21 14:41:42.906901 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.906868 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-967cb\" (UniqueName: \"kubernetes.io/projected/ed7f3c77-cdc2-4f59-939c-1bb0703ee060-kube-api-access-967cb\") pod \"maas-api-5c64fc8cbc-w9f4j\" (UID: \"ed7f3c77-cdc2-4f59-939c-1bb0703ee060\") " pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:42.907096 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:42.906928 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/ed7f3c77-cdc2-4f59-939c-1bb0703ee060-maas-api-tls\") pod \"maas-api-5c64fc8cbc-w9f4j\" (UID: \"ed7f3c77-cdc2-4f59-939c-1bb0703ee060\") " pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:43.007556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:43.007457 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-967cb\" (UniqueName: \"kubernetes.io/projected/ed7f3c77-cdc2-4f59-939c-1bb0703ee060-kube-api-access-967cb\") pod \"maas-api-5c64fc8cbc-w9f4j\" (UID: \"ed7f3c77-cdc2-4f59-939c-1bb0703ee060\") " pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:43.007556 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:43.007519 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/ed7f3c77-cdc2-4f59-939c-1bb0703ee060-maas-api-tls\") pod \"maas-api-5c64fc8cbc-w9f4j\" (UID: \"ed7f3c77-cdc2-4f59-939c-1bb0703ee060\") " pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:43.010166 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:43.010143 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/ed7f3c77-cdc2-4f59-939c-1bb0703ee060-maas-api-tls\") pod \"maas-api-5c64fc8cbc-w9f4j\" (UID: \"ed7f3c77-cdc2-4f59-939c-1bb0703ee060\") " pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:43.018477 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:43.018439 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-967cb\" (UniqueName: \"kubernetes.io/projected/ed7f3c77-cdc2-4f59-939c-1bb0703ee060-kube-api-access-967cb\") pod \"maas-api-5c64fc8cbc-w9f4j\" (UID: \"ed7f3c77-cdc2-4f59-939c-1bb0703ee060\") " pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:43.054316 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:43.054276 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:43.390181 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:43.390150 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-api-5c64fc8cbc-w9f4j"] Apr 21 14:41:43.391655 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:41:43.391621 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded7f3c77_cdc2_4f59_939c_1bb0703ee060.slice/crio-0931ae4689054edcd7bc3214fb793e96b1e7884aedf5c5b19509851c8bf88095 WatchSource:0}: Error finding container 0931ae4689054edcd7bc3214fb793e96b1e7884aedf5c5b19509851c8bf88095: Status 404 returned error can't find the container with id 0931ae4689054edcd7bc3214fb793e96b1e7884aedf5c5b19509851c8bf88095 Apr 21 14:41:44.302031 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:44.301992 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" event={"ID":"ed7f3c77-cdc2-4f59-939c-1bb0703ee060","Type":"ContainerStarted","Data":"0931ae4689054edcd7bc3214fb793e96b1e7884aedf5c5b19509851c8bf88095"} Apr 21 14:41:46.311350 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:46.311311 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" event={"ID":"ed7f3c77-cdc2-4f59-939c-1bb0703ee060","Type":"ContainerStarted","Data":"cd07c6e71a95d84cd77acc3223bdff479dfd26bd05a630cf5e7223e24f4bddc0"} Apr 21 14:41:46.311800 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:46.311477 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:41:46.330177 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:46.330120 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" podStartSLOduration=2.279067322 podStartE2EDuration="4.330104972s" podCreationTimestamp="2026-04-21 14:41:42 +0000 UTC" firstStartedPulling="2026-04-21 14:41:43.393451377 +0000 UTC m=+936.850935862" lastFinishedPulling="2026-04-21 14:41:45.444489024 +0000 UTC m=+938.901973512" observedRunningTime="2026-04-21 14:41:46.327858714 +0000 UTC m=+939.785343221" watchObservedRunningTime="2026-04-21 14:41:46.330104972 +0000 UTC m=+939.787589478" Apr 21 14:41:52.320509 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:41:52.320481 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-api-5c64fc8cbc-w9f4j" Apr 21 14:42:17.155396 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.155136 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-7ff78486fb-rs9pn"] Apr 21 14:42:17.161266 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.161244 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.165355 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.165334 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"authorino-oidc-ca\"" Apr 21 14:42:17.166110 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.166089 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-7ff78486fb-rs9pn"] Apr 21 14:42:17.201463 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.201435 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knxhg\" (UniqueName: \"kubernetes.io/projected/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-kube-api-access-knxhg\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.201648 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.201491 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-oidc-ca\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.201648 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.201618 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-tls-cert\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.303038 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.302992 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-oidc-ca\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.303237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.303061 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-tls-cert\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.303237 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.303105 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-knxhg\" (UniqueName: \"kubernetes.io/projected/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-kube-api-access-knxhg\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.303712 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.303690 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-oidc-ca\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.305761 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.305738 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-tls-cert\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.313501 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.313476 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-knxhg\" (UniqueName: \"kubernetes.io/projected/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-kube-api-access-knxhg\") pod \"authorino-7ff78486fb-rs9pn\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.471757 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.471711 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:42:17.616672 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.616648 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-7ff78486fb-rs9pn"] Apr 21 14:42:17.618678 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:42:17.618646 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8dccfb6_49d6_4c26_ba51_69a9dfce4ffd.slice/crio-f8c788b8ecf392c440ba448deb21b3ffa8e19cb02f60139507076e91fd2a9b24 WatchSource:0}: Error finding container f8c788b8ecf392c440ba448deb21b3ffa8e19cb02f60139507076e91fd2a9b24: Status 404 returned error can't find the container with id f8c788b8ecf392c440ba448deb21b3ffa8e19cb02f60139507076e91fd2a9b24 Apr 21 14:42:17.619859 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:17.619840 2580 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 14:42:18.426966 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.426927 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" event={"ID":"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd","Type":"ContainerStarted","Data":"409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e"} Apr 21 14:42:18.427396 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.426975 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" event={"ID":"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd","Type":"ContainerStarted","Data":"f8c788b8ecf392c440ba448deb21b3ffa8e19cb02f60139507076e91fd2a9b24"} Apr 21 14:42:18.452772 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.452703 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" podStartSLOduration=0.990464387 podStartE2EDuration="1.452686457s" podCreationTimestamp="2026-04-21 14:42:17 +0000 UTC" firstStartedPulling="2026-04-21 14:42:17.619996104 +0000 UTC m=+971.077480588" lastFinishedPulling="2026-04-21 14:42:18.082218166 +0000 UTC m=+971.539702658" observedRunningTime="2026-04-21 14:42:18.450177217 +0000 UTC m=+971.907661723" watchObservedRunningTime="2026-04-21 14:42:18.452686457 +0000 UTC m=+971.910170964" Apr 21 14:42:18.482689 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.482656 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-b865f9f7b-ls2zr"] Apr 21 14:42:18.483589 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.482928 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" podUID="97a60321-437f-484d-a535-ca02c03d8b11" containerName="authorino" containerID="cri-o://ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24" gracePeriod=30 Apr 21 14:42:18.737580 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.737534 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:42:18.814593 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.814533 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2x9b\" (UniqueName: \"kubernetes.io/projected/97a60321-437f-484d-a535-ca02c03d8b11-kube-api-access-d2x9b\") pod \"97a60321-437f-484d-a535-ca02c03d8b11\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " Apr 21 14:42:18.814784 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.814657 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/97a60321-437f-484d-a535-ca02c03d8b11-tls-cert\") pod \"97a60321-437f-484d-a535-ca02c03d8b11\" (UID: \"97a60321-437f-484d-a535-ca02c03d8b11\") " Apr 21 14:42:18.816884 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.816853 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a60321-437f-484d-a535-ca02c03d8b11-kube-api-access-d2x9b" (OuterVolumeSpecName: "kube-api-access-d2x9b") pod "97a60321-437f-484d-a535-ca02c03d8b11" (UID: "97a60321-437f-484d-a535-ca02c03d8b11"). InnerVolumeSpecName "kube-api-access-d2x9b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:42:18.826343 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.826311 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a60321-437f-484d-a535-ca02c03d8b11-tls-cert" (OuterVolumeSpecName: "tls-cert") pod "97a60321-437f-484d-a535-ca02c03d8b11" (UID: "97a60321-437f-484d-a535-ca02c03d8b11"). InnerVolumeSpecName "tls-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:42:18.916170 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.916130 2580 reconciler_common.go:299] "Volume detached for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/97a60321-437f-484d-a535-ca02c03d8b11-tls-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:42:18.916170 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:18.916163 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d2x9b\" (UniqueName: \"kubernetes.io/projected/97a60321-437f-484d-a535-ca02c03d8b11-kube-api-access-d2x9b\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:42:19.431580 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.431522 2580 generic.go:358] "Generic (PLEG): container finished" podID="97a60321-437f-484d-a535-ca02c03d8b11" containerID="ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24" exitCode=0 Apr 21 14:42:19.432032 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.431596 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" Apr 21 14:42:19.432032 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.431599 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" event={"ID":"97a60321-437f-484d-a535-ca02c03d8b11","Type":"ContainerDied","Data":"ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24"} Apr 21 14:42:19.432032 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.431641 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-b865f9f7b-ls2zr" event={"ID":"97a60321-437f-484d-a535-ca02c03d8b11","Type":"ContainerDied","Data":"1a346d5072a4a7924b8b23c14a02251518c0dfb30349a8a77bc93371d2a91abe"} Apr 21 14:42:19.432032 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.431661 2580 scope.go:117] "RemoveContainer" containerID="ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24" Apr 21 14:42:19.440311 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.440289 2580 scope.go:117] "RemoveContainer" containerID="ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24" Apr 21 14:42:19.440555 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:42:19.440536 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24\": container with ID starting with ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24 not found: ID does not exist" containerID="ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24" Apr 21 14:42:19.440661 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.440581 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24"} err="failed to get container status \"ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24\": rpc error: code = NotFound desc = could not find container \"ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24\": container with ID starting with ff4e6bcaea75e841d387bb2a18cce1e7a385bea0dbf44c104d43f6882a12ec24 not found: ID does not exist" Apr 21 14:42:19.457109 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.457084 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-b865f9f7b-ls2zr"] Apr 21 14:42:19.459172 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:19.459151 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-b865f9f7b-ls2zr"] Apr 21 14:42:21.245765 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:21.245684 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a60321-437f-484d-a535-ca02c03d8b11" path="/var/lib/kubelet/pods/97a60321-437f-484d-a535-ca02c03d8b11/volumes" Apr 21 14:42:27.199888 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:27.199845 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:42:45.300225 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:45.300189 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:42:52.127819 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:42:52.127773 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:43:02.298384 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:02.295957 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:43:20.204519 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:20.204482 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:43:26.991346 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:26.991302 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:43:59.565701 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.565602 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-fddf965c7-cjhz8"] Apr 21 14:43:59.566164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.565970 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="97a60321-437f-484d-a535-ca02c03d8b11" containerName="authorino" Apr 21 14:43:59.566164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.565983 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a60321-437f-484d-a535-ca02c03d8b11" containerName="authorino" Apr 21 14:43:59.566164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.566051 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="97a60321-437f-484d-a535-ca02c03d8b11" containerName="authorino" Apr 21 14:43:59.568980 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.568961 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.578384 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.578361 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-fddf965c7-cjhz8"] Apr 21 14:43:59.629058 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.629019 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7m99\" (UniqueName: \"kubernetes.io/projected/65d5e11c-46c9-4253-a6d0-e2e79c96d589-kube-api-access-q7m99\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.629233 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.629066 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/65d5e11c-46c9-4253-a6d0-e2e79c96d589-tls-cert\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.629233 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.629097 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/65d5e11c-46c9-4253-a6d0-e2e79c96d589-oidc-ca\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.730112 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.730078 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-q7m99\" (UniqueName: \"kubernetes.io/projected/65d5e11c-46c9-4253-a6d0-e2e79c96d589-kube-api-access-q7m99\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.730281 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.730121 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/65d5e11c-46c9-4253-a6d0-e2e79c96d589-tls-cert\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.730281 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.730159 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/65d5e11c-46c9-4253-a6d0-e2e79c96d589-oidc-ca\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.730911 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.730892 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/65d5e11c-46c9-4253-a6d0-e2e79c96d589-oidc-ca\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.732806 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.732787 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/65d5e11c-46c9-4253-a6d0-e2e79c96d589-tls-cert\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.738717 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.738695 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7m99\" (UniqueName: \"kubernetes.io/projected/65d5e11c-46c9-4253-a6d0-e2e79c96d589-kube-api-access-q7m99\") pod \"authorino-fddf965c7-cjhz8\" (UID: \"65d5e11c-46c9-4253-a6d0-e2e79c96d589\") " pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:43:59.879256 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:43:59.879166 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-fddf965c7-cjhz8" Apr 21 14:44:00.008337 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:00.008307 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-fddf965c7-cjhz8"] Apr 21 14:44:00.010624 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:44:00.010594 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65d5e11c_46c9_4253_a6d0_e2e79c96d589.slice/crio-ea2dc3f8f11fe690a1947397d61750f357c8bf7a342383016a923171cdf5a1ca WatchSource:0}: Error finding container ea2dc3f8f11fe690a1947397d61750f357c8bf7a342383016a923171cdf5a1ca: Status 404 returned error can't find the container with id ea2dc3f8f11fe690a1947397d61750f357c8bf7a342383016a923171cdf5a1ca Apr 21 14:44:00.783663 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:00.783613 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-fddf965c7-cjhz8" event={"ID":"65d5e11c-46c9-4253-a6d0-e2e79c96d589","Type":"ContainerStarted","Data":"e0697b76ae2dc35198b00a88d2bc4bafe2b563471288bcf5b190a1a7f8ba1740"} Apr 21 14:44:00.783663 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:00.783662 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-fddf965c7-cjhz8" event={"ID":"65d5e11c-46c9-4253-a6d0-e2e79c96d589","Type":"ContainerStarted","Data":"ea2dc3f8f11fe690a1947397d61750f357c8bf7a342383016a923171cdf5a1ca"} Apr 21 14:44:00.843700 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:00.843635 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-fddf965c7-cjhz8" podStartSLOduration=1.4964806099999999 podStartE2EDuration="1.843615889s" podCreationTimestamp="2026-04-21 14:43:59 +0000 UTC" firstStartedPulling="2026-04-21 14:44:00.011823197 +0000 UTC m=+1073.469307681" lastFinishedPulling="2026-04-21 14:44:00.358958463 +0000 UTC m=+1073.816442960" observedRunningTime="2026-04-21 14:44:00.80512516 +0000 UTC m=+1074.262609668" watchObservedRunningTime="2026-04-21 14:44:00.843615889 +0000 UTC m=+1074.301100398" Apr 21 14:44:00.845718 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:00.845391 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-7ff78486fb-rs9pn"] Apr 21 14:44:00.845718 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:00.845670 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" podUID="c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" containerName="authorino" containerID="cri-o://409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e" gracePeriod=30 Apr 21 14:44:01.107460 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.107438 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:44:01.242917 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.242883 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-oidc-ca\") pod \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " Apr 21 14:44:01.243123 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.242967 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knxhg\" (UniqueName: \"kubernetes.io/projected/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-kube-api-access-knxhg\") pod \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " Apr 21 14:44:01.243123 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.242985 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-tls-cert\") pod \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\" (UID: \"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd\") " Apr 21 14:44:01.245264 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.245232 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-kube-api-access-knxhg" (OuterVolumeSpecName: "kube-api-access-knxhg") pod "c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" (UID: "c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd"). InnerVolumeSpecName "kube-api-access-knxhg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:44:01.249094 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.249066 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-oidc-ca" (OuterVolumeSpecName: "oidc-ca") pod "c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" (UID: "c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd"). InnerVolumeSpecName "oidc-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 14:44:01.253950 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.253918 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-tls-cert" (OuterVolumeSpecName: "tls-cert") pod "c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" (UID: "c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd"). InnerVolumeSpecName "tls-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 14:44:01.344131 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.344044 2580 reconciler_common.go:299] "Volume detached for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-oidc-ca\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:44:01.344131 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.344076 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-knxhg\" (UniqueName: \"kubernetes.io/projected/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-kube-api-access-knxhg\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:44:01.344131 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.344087 2580 reconciler_common.go:299] "Volume detached for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd-tls-cert\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:44:01.787976 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.787943 2580 generic.go:358] "Generic (PLEG): container finished" podID="c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" containerID="409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e" exitCode=0 Apr 21 14:44:01.788422 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.788010 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" Apr 21 14:44:01.788422 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.788035 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" event={"ID":"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd","Type":"ContainerDied","Data":"409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e"} Apr 21 14:44:01.788422 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.788075 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7ff78486fb-rs9pn" event={"ID":"c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd","Type":"ContainerDied","Data":"f8c788b8ecf392c440ba448deb21b3ffa8e19cb02f60139507076e91fd2a9b24"} Apr 21 14:44:01.788422 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.788092 2580 scope.go:117] "RemoveContainer" containerID="409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e" Apr 21 14:44:01.796630 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.796607 2580 scope.go:117] "RemoveContainer" containerID="409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e" Apr 21 14:44:01.796888 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:44:01.796872 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e\": container with ID starting with 409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e not found: ID does not exist" containerID="409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e" Apr 21 14:44:01.796935 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.796898 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e"} err="failed to get container status \"409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e\": rpc error: code = NotFound desc = could not find container \"409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e\": container with ID starting with 409930ba40a0fe82f2b5f3c38f517aadad3d60f1c3b82814b855e6bdf60dab6e not found: ID does not exist" Apr 21 14:44:01.809041 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.809012 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-7ff78486fb-rs9pn"] Apr 21 14:44:01.813483 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:01.813453 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-7ff78486fb-rs9pn"] Apr 21 14:44:03.246628 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:03.246592 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" path="/var/lib/kubelet/pods/c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd/volumes" Apr 21 14:44:32.696816 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:32.696779 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:44:43.786678 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:43.786637 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:44:51.888128 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:44:51.888091 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:45:02.989130 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:02.989091 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:45:11.286930 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:11.286886 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:45:22.892325 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:22.892247 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:45:29.031173 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:29.031136 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-744d89cb86-bkwjx"] Apr 21 14:45:29.031609 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:29.031381 2580 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-controller-744d89cb86-bkwjx" podUID="7adf31a4-3dfc-451d-b65d-66d6d8ff93d7" containerName="manager" containerID="cri-o://be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c" gracePeriod=10 Apr 21 14:45:29.290271 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:29.290205 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:45:29.352152 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:29.352122 2580 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfh7t\" (UniqueName: \"kubernetes.io/projected/7adf31a4-3dfc-451d-b65d-66d6d8ff93d7-kube-api-access-nfh7t\") pod \"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7\" (UID: \"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7\") " Apr 21 14:45:29.354359 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:29.354324 2580 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7adf31a4-3dfc-451d-b65d-66d6d8ff93d7-kube-api-access-nfh7t" (OuterVolumeSpecName: "kube-api-access-nfh7t") pod "7adf31a4-3dfc-451d-b65d-66d6d8ff93d7" (UID: "7adf31a4-3dfc-451d-b65d-66d6d8ff93d7"). InnerVolumeSpecName "kube-api-access-nfh7t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 14:45:29.453547 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:29.453504 2580 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nfh7t\" (UniqueName: \"kubernetes.io/projected/7adf31a4-3dfc-451d-b65d-66d6d8ff93d7-kube-api-access-nfh7t\") on node \"ip-10-0-141-61.ec2.internal\" DevicePath \"\"" Apr 21 14:45:30.100115 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.100079 2580 generic.go:358] "Generic (PLEG): container finished" podID="7adf31a4-3dfc-451d-b65d-66d6d8ff93d7" containerID="be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c" exitCode=0 Apr 21 14:45:30.100634 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.100135 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-744d89cb86-bkwjx" event={"ID":"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7","Type":"ContainerDied","Data":"be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c"} Apr 21 14:45:30.100634 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.100161 2580 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-744d89cb86-bkwjx" Apr 21 14:45:30.100634 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.100168 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-744d89cb86-bkwjx" event={"ID":"7adf31a4-3dfc-451d-b65d-66d6d8ff93d7","Type":"ContainerDied","Data":"b8d5ba90d2a70fc71de04ae8686cf132a55e7d9ddbcfaa87fcb521fd62bc1207"} Apr 21 14:45:30.100634 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.100189 2580 scope.go:117] "RemoveContainer" containerID="be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c" Apr 21 14:45:30.109013 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.108996 2580 scope.go:117] "RemoveContainer" containerID="be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c" Apr 21 14:45:30.109307 ip-10-0-141-61 kubenswrapper[2580]: E0421 14:45:30.109289 2580 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c\": container with ID starting with be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c not found: ID does not exist" containerID="be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c" Apr 21 14:45:30.109363 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.109315 2580 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c"} err="failed to get container status \"be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c\": rpc error: code = NotFound desc = could not find container \"be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c\": container with ID starting with be54dd4bd0b627fa2fe410bd4a5943a24f62be2b849d5fac16a5be9cc470333c not found: ID does not exist" Apr 21 14:45:30.121029 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.121000 2580 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-744d89cb86-bkwjx"] Apr 21 14:45:30.127893 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.127870 2580 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-744d89cb86-bkwjx"] Apr 21 14:45:30.901528 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.901496 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-744d89cb86-bjwjg"] Apr 21 14:45:30.901892 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.901879 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7adf31a4-3dfc-451d-b65d-66d6d8ff93d7" containerName="manager" Apr 21 14:45:30.901939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.901895 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="7adf31a4-3dfc-451d-b65d-66d6d8ff93d7" containerName="manager" Apr 21 14:45:30.901939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.901904 2580 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" containerName="authorino" Apr 21 14:45:30.901939 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.901911 2580 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" containerName="authorino" Apr 21 14:45:30.902028 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.901967 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="7adf31a4-3dfc-451d-b65d-66d6d8ff93d7" containerName="manager" Apr 21 14:45:30.902028 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.901976 2580 memory_manager.go:356] "RemoveStaleState removing state" podUID="c8dccfb6-49d6-4c26-ba51-69a9dfce4ffd" containerName="authorino" Apr 21 14:45:30.906264 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.906246 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-744d89cb86-bjwjg" Apr 21 14:45:30.909242 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.909220 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-controller-dockercfg-mqm8t\"" Apr 21 14:45:30.916319 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.916292 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-744d89cb86-bjwjg"] Apr 21 14:45:30.966262 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:30.966234 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp5bb\" (UniqueName: \"kubernetes.io/projected/b7ed0f19-f795-4675-a1d4-890a5a77d842-kube-api-access-rp5bb\") pod \"maas-controller-744d89cb86-bjwjg\" (UID: \"b7ed0f19-f795-4675-a1d4-890a5a77d842\") " pod="opendatahub/maas-controller-744d89cb86-bjwjg" Apr 21 14:45:31.066815 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:31.066776 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rp5bb\" (UniqueName: \"kubernetes.io/projected/b7ed0f19-f795-4675-a1d4-890a5a77d842-kube-api-access-rp5bb\") pod \"maas-controller-744d89cb86-bjwjg\" (UID: \"b7ed0f19-f795-4675-a1d4-890a5a77d842\") " pod="opendatahub/maas-controller-744d89cb86-bjwjg" Apr 21 14:45:31.075175 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:31.075150 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp5bb\" (UniqueName: \"kubernetes.io/projected/b7ed0f19-f795-4675-a1d4-890a5a77d842-kube-api-access-rp5bb\") pod \"maas-controller-744d89cb86-bjwjg\" (UID: \"b7ed0f19-f795-4675-a1d4-890a5a77d842\") " pod="opendatahub/maas-controller-744d89cb86-bjwjg" Apr 21 14:45:31.218067 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:31.218037 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-744d89cb86-bjwjg" Apr 21 14:45:31.245581 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:31.245536 2580 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7adf31a4-3dfc-451d-b65d-66d6d8ff93d7" path="/var/lib/kubelet/pods/7adf31a4-3dfc-451d-b65d-66d6d8ff93d7/volumes" Apr 21 14:45:31.346254 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:31.346232 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-744d89cb86-bjwjg"] Apr 21 14:45:31.348672 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:45:31.348635 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7ed0f19_f795_4675_a1d4_890a5a77d842.slice/crio-ba10ed51e436a94ed15467c61667af8980a9e6dc0116a99962ad7b5464fc2556 WatchSource:0}: Error finding container ba10ed51e436a94ed15467c61667af8980a9e6dc0116a99962ad7b5464fc2556: Status 404 returned error can't find the container with id ba10ed51e436a94ed15467c61667af8980a9e6dc0116a99962ad7b5464fc2556 Apr 21 14:45:32.109792 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:32.109752 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-744d89cb86-bjwjg" event={"ID":"b7ed0f19-f795-4675-a1d4-890a5a77d842","Type":"ContainerStarted","Data":"a5152617ac0dc8ab0cebbb577413da50fb7316f884befa44ea7b68bbb63bdc62"} Apr 21 14:45:32.109989 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:32.109799 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-744d89cb86-bjwjg" event={"ID":"b7ed0f19-f795-4675-a1d4-890a5a77d842","Type":"ContainerStarted","Data":"ba10ed51e436a94ed15467c61667af8980a9e6dc0116a99962ad7b5464fc2556"} Apr 21 14:45:32.109989 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:32.109832 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-744d89cb86-bjwjg" Apr 21 14:45:32.133331 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:32.133280 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-744d89cb86-bjwjg" podStartSLOduration=1.535631011 podStartE2EDuration="2.133265216s" podCreationTimestamp="2026-04-21 14:45:30 +0000 UTC" firstStartedPulling="2026-04-21 14:45:31.350321222 +0000 UTC m=+1164.807805707" lastFinishedPulling="2026-04-21 14:45:31.947955424 +0000 UTC m=+1165.405439912" observedRunningTime="2026-04-21 14:45:32.132000574 +0000 UTC m=+1165.589485094" watchObservedRunningTime="2026-04-21 14:45:32.133265216 +0000 UTC m=+1165.590749719" Apr 21 14:45:43.120136 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:43.120095 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-controller-744d89cb86-bjwjg" Apr 21 14:45:53.707271 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:53.707232 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:45:57.986745 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:45:57.986710 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:46:07.182291 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:07.182260 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:46:07.183492 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:07.183471 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:46:07.183686 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:07.183665 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:46:07.185052 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:07.185025 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:46:07.186674 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:07.186657 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:46:07.188035 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:07.188018 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:46:26.686523 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:26.686485 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:46:41.604263 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:46:41.604226 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:47:20.887300 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:47:20.887268 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:47:37.501082 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:47:37.501043 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:47:52.760093 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:47:52.760057 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:48:09.292852 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:48:09.292815 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:49:03.015334 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:49:03.015293 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:49:12.593612 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:49:12.593560 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:49:28.396186 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:49:28.396154 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:49:38.283670 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:49:38.283622 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:49:53.993151 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:49:53.992928 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:50:03.012468 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:50:03.012426 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:50:34.914891 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:50:34.914857 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:50:44.298387 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:50:44.298353 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:50:52.504473 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:50:52.504443 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:51:01.102187 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:01.102140 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:51:07.208036 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:07.208008 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:51:07.209592 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:07.209555 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:51:07.210696 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:07.210671 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:51:07.212251 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:07.212232 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:51:07.212778 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:07.212762 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:51:07.215283 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:07.215263 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:51:09.601634 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:09.601597 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:51:17.005464 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:17.005425 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:51:21.895002 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:21.894922 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:51:26.196955 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:26.196919 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:51:38.899140 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:51:38.899096 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:52:27.317990 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:52:27.317946 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:52:36.100687 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:52:36.100650 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:52:44.694302 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:52:44.694262 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:52:54.006807 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:52:54.006720 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:53:01.994795 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:53:01.994754 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:53:10.694734 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:53:10.694700 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:53:19.990994 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:53:19.990956 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:53:28.395386 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:53:28.395351 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:53:37.724147 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:53:37.724107 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:53:46.192979 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:53:46.192939 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:53:55.708030 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:53:55.707990 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:54:03.613826 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:54:03.613792 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:54:12.891978 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:54:12.891940 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:54:20.904533 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:54:20.904433 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:54:29.594254 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:54:29.594217 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:54:38.405776 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:54:38.405737 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:54:46.803086 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:54:46.803050 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:54:55.198811 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:54:55.198766 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:56:07.232429 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:07.232392 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:56:07.233876 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:07.233855 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:56:07.234088 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:07.234069 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 14:56:07.235692 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:07.235675 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 14:56:07.236696 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:07.236675 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:56:07.238826 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:07.238808 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 14:56:53.144889 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.144848 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg"] Apr 21 14:56:53.148309 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.148284 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.148705 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.148675 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fn9v\" (UniqueName: \"kubernetes.io/projected/0332eb41-94f2-45d2-ab55-4e1436fe2e9a-kube-api-access-5fn9v\") pod \"kuadrant-operator-controller-manager-55c7f4c975-w8rsg\" (UID: \"0332eb41-94f2-45d2-ab55-4e1436fe2e9a\") " pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.148777 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.148731 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/0332eb41-94f2-45d2-ab55-4e1436fe2e9a-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-55c7f4c975-w8rsg\" (UID: \"0332eb41-94f2-45d2-ab55-4e1436fe2e9a\") " pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.153706 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.153691 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"kuadrant-operator-controller-manager-dockercfg-75flw\"" Apr 21 14:56:53.168293 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.168271 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg"] Apr 21 14:56:53.249164 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.249133 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5fn9v\" (UniqueName: \"kubernetes.io/projected/0332eb41-94f2-45d2-ab55-4e1436fe2e9a-kube-api-access-5fn9v\") pod \"kuadrant-operator-controller-manager-55c7f4c975-w8rsg\" (UID: \"0332eb41-94f2-45d2-ab55-4e1436fe2e9a\") " pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.249335 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.249179 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/0332eb41-94f2-45d2-ab55-4e1436fe2e9a-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-55c7f4c975-w8rsg\" (UID: \"0332eb41-94f2-45d2-ab55-4e1436fe2e9a\") " pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.249601 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.249556 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"extensions-socket-volume\" (UniqueName: \"kubernetes.io/empty-dir/0332eb41-94f2-45d2-ab55-4e1436fe2e9a-extensions-socket-volume\") pod \"kuadrant-operator-controller-manager-55c7f4c975-w8rsg\" (UID: \"0332eb41-94f2-45d2-ab55-4e1436fe2e9a\") " pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.260860 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.260839 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fn9v\" (UniqueName: \"kubernetes.io/projected/0332eb41-94f2-45d2-ab55-4e1436fe2e9a-kube-api-access-5fn9v\") pod \"kuadrant-operator-controller-manager-55c7f4c975-w8rsg\" (UID: \"0332eb41-94f2-45d2-ab55-4e1436fe2e9a\") " pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.458378 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.458352 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:53.591093 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.591033 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg"] Apr 21 14:56:53.593409 ip-10-0-141-61 kubenswrapper[2580]: W0421 14:56:53.593377 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0332eb41_94f2_45d2_ab55_4e1436fe2e9a.slice/crio-306ba965691cdd82e5beaa311af4754f36634e4706b3efc47185aa6c440af089 WatchSource:0}: Error finding container 306ba965691cdd82e5beaa311af4754f36634e4706b3efc47185aa6c440af089: Status 404 returned error can't find the container with id 306ba965691cdd82e5beaa311af4754f36634e4706b3efc47185aa6c440af089 Apr 21 14:56:53.595771 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:53.595756 2580 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 14:56:54.462756 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:54.462711 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" event={"ID":"0332eb41-94f2-45d2-ab55-4e1436fe2e9a","Type":"ContainerStarted","Data":"54219eee90f242f45307414de07d4a958bf63f2d92f348770c66165c1d7ade0d"} Apr 21 14:56:54.462756 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:54.462752 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" event={"ID":"0332eb41-94f2-45d2-ab55-4e1436fe2e9a","Type":"ContainerStarted","Data":"306ba965691cdd82e5beaa311af4754f36634e4706b3efc47185aa6c440af089"} Apr 21 14:56:54.463319 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:54.462878 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:56:54.488012 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:56:54.487963 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" podStartSLOduration=1.487947065 podStartE2EDuration="1.487947065s" podCreationTimestamp="2026-04-21 14:56:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 14:56:54.484667601 +0000 UTC m=+1847.942152119" watchObservedRunningTime="2026-04-21 14:56:54.487947065 +0000 UTC m=+1847.945431574" Apr 21 14:57:05.468670 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:57:05.468633 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/kuadrant-operator-controller-manager-55c7f4c975-w8rsg" Apr 21 14:57:12.932426 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:57:12.932389 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:57:18.325647 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:57:18.325607 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:57:44.148005 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:57:44.147965 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:57:51.522270 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:57:51.522223 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:58:00.021027 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:58:00.020991 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:58:10.721901 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:58:10.721863 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:58:19.844908 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:58:19.844871 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:58:30.923836 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:58:30.923802 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:58:38.834296 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:58:38.834260 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:58:49.321789 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:58:49.321755 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:58:58.225087 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:58:58.224989 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:59:08.174391 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:59:08.174351 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:59:17.747324 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:59:17.747283 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 14:59:51.029431 ip-10-0-141-61 kubenswrapper[2580]: I0421 14:59:51.029393 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:00:33.038082 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:00:33.038048 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:00:41.815046 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:00:41.815011 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:00:50.825751 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:00:50.825713 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:00:59.630897 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:00:59.630856 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:01:07.257842 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:07.257805 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 15:01:07.259358 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:07.259338 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 15:01:07.261297 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:07.261281 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 15:01:07.262299 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:07.262280 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 15:01:07.262791 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:07.262768 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 15:01:07.266281 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:07.266260 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 15:01:08.126392 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:08.126353 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:01:20.517178 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:20.517143 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:01:30.024797 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:30.024761 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:01:36.322934 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:36.322899 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:01:46.196699 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:46.196661 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:01:54.729888 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:01:54.729800 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:02:02.913423 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:02:02.913388 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:02:13.519866 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:02:13.519827 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:02:31.018906 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:02:31.018870 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:02:39.223087 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:02:39.223050 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:02:47.616862 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:02:47.616822 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:02:55.719999 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:02:55.719958 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:03:13.522261 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:03:13.522224 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:03:21.629052 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:03:21.628966 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:03:31.625134 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:03:31.625096 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:03:38.325147 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:03:38.325110 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:03:47.944299 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:03:47.944263 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:03:56.636897 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:03:56.636856 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:04:05.451040 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:04:05.451001 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:04:18.532616 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:04:18.532559 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:04:27.219677 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:04:27.219642 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:04:38.936979 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:04:38.936934 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:04:47.630797 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:04:47.630759 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:04:59.436868 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:04:59.436785 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:05:09.140399 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:05:09.140360 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:05:15.752098 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:05:15.752061 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:05:33.330353 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:05:33.330316 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:05:42.366075 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:05:42.366037 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:05:50.395275 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:05:50.395233 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:05:58.872839 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:05:58.872803 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:06:07.283937 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:07.283908 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 15:06:07.285496 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:07.285472 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 15:06:07.287160 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:07.287139 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 15:06:07.288677 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:07.288656 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 15:06:07.288769 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:07.288702 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 15:06:07.291718 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:07.291697 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ts88z_05386859-dd8b-49f4-8271-44949e102c17/ovn-acl-logging/0.log" Apr 21 15:06:23.054801 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:23.054713 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:06:34.835626 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:06:34.835586 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-78c99df468-xls6n"] Apr 21 15:10:20.727454 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:20.727395 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-fddf965c7-cjhz8_65d5e11c-46c9-4253-a6d0-e2e79c96d589/authorino/0.log" Apr 21 15:10:24.618616 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:24.618581 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_maas-api-5c64fc8cbc-w9f4j_ed7f3c77-cdc2-4f59-939c-1bb0703ee060/maas-api/0.log" Apr 21 15:10:24.735158 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:24.735124 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_maas-controller-744d89cb86-bjwjg_b7ed0f19-f795-4675-a1d4-890a5a77d842/manager/0.log" Apr 21 15:10:24.969059 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:24.969018 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_opendatahub-operator-controller-manager-7df645bd74-9mxmk_fbebfd28-f8f1-40d8-83f8-99bbbb82a630/manager/0.log" Apr 21 15:10:26.582891 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:26.582853 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-fddf965c7-cjhz8_65d5e11c-46c9-4253-a6d0-e2e79c96d589/authorino/0.log" Apr 21 15:10:26.704114 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:26.704079 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-operator-657f44b778-kpwkq_dd86a861-3beb-495b-b60e-6361f51519cc/manager/0.log" Apr 21 15:10:26.807847 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:26.807811 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_dns-operator-controller-manager-648d5c98bc-xn8rl_95a16468-5acb-42be-a1f1-8ffaf3abf10a/manager/0.log" Apr 21 15:10:27.025814 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:27.025784 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_kuadrant-operator-catalog-6q598_856e4d1a-8dc6-487a-848b-2f081de3c7ff/registry-server/0.log" Apr 21 15:10:27.143805 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:27.143770 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_kuadrant-operator-controller-manager-55c7f4c975-w8rsg_0332eb41-94f2-45d2-ab55-4e1436fe2e9a/manager/0.log" Apr 21 15:10:27.250719 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:27.250686 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_limitador-limitador-78c99df468-xls6n_8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d/limitador/0.log" Apr 21 15:10:27.696264 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:27.696216 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d_37d51f19-e7f6-4871-8cb3-289134a6eabb/istio-proxy/0.log" Apr 21 15:10:28.163683 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:28.163603 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_maas-default-gateway-openshift-default-58b6f876-vbvzt_9017a28a-4013-4114-b8fb-4d3612657389/istio-proxy/0.log" Apr 21 15:10:28.272465 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:28.272439 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-6d6c54d4dd-mb7h8_0284ee69-97a8-4c33-98ee-1bb75e7f461b/router/0.log" Apr 21 15:10:36.541927 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:36.541895 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-ztmwm_ac750e76-e718-4228-a16a-ea303bb3ed02/global-pull-secret-syncer/0.log" Apr 21 15:10:36.678751 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:36.678724 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-jzct4_17883801-bbff-4ed7-8ba5-44081b8c5476/konnectivity-agent/0.log" Apr 21 15:10:36.756990 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:36.756962 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-141-61.ec2.internal_30f945062f07b4b1388971557d0103c9/haproxy/0.log" Apr 21 15:10:41.539758 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:41.539726 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-fddf965c7-cjhz8_65d5e11c-46c9-4253-a6d0-e2e79c96d589/authorino/0.log" Apr 21 15:10:41.578337 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:41.578313 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-operator-657f44b778-kpwkq_dd86a861-3beb-495b-b60e-6361f51519cc/manager/0.log" Apr 21 15:10:41.638898 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:41.638864 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_dns-operator-controller-manager-648d5c98bc-xn8rl_95a16468-5acb-42be-a1f1-8ffaf3abf10a/manager/0.log" Apr 21 15:10:41.750885 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:41.750855 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_kuadrant-operator-catalog-6q598_856e4d1a-8dc6-487a-848b-2f081de3c7ff/registry-server/0.log" Apr 21 15:10:41.821099 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:41.821018 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_kuadrant-operator-controller-manager-55c7f4c975-w8rsg_0332eb41-94f2-45d2-ab55-4e1436fe2e9a/manager/0.log" Apr 21 15:10:41.846956 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:41.846928 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_limitador-limitador-78c99df468-xls6n_8b20ed5b-ca5f-41d5-aad2-1645e8f2d07d/limitador/0.log" Apr 21 15:10:43.667118 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:43.667077 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/1.log" Apr 21 15:10:43.750083 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:43.750042 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-pn2df_f355a71a-6279-4fa0-9123-6a056ee88f72/cluster-monitoring-operator/0.log" Apr 21 15:10:43.909531 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:43.909503 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_monitoring-plugin-7dccd58f55-fk6qh_1779e5fc-59fd-4957-a145-e08e128869c7/monitoring-plugin/0.log" Apr 21 15:10:44.238024 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:44.238000 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-ndjgn_48f962b2-2b0a-455d-91f8-1a90cdb8f2db/node-exporter/0.log" Apr 21 15:10:44.282395 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:44.282365 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-ndjgn_48f962b2-2b0a-455d-91f8-1a90cdb8f2db/kube-rbac-proxy/0.log" Apr 21 15:10:44.334346 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:44.334323 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-ndjgn_48f962b2-2b0a-455d-91f8-1a90cdb8f2db/init-textfile/0.log" Apr 21 15:10:45.407101 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.407069 2580 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn"] Apr 21 15:10:45.410448 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.410432 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.413305 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.413276 2580 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-qvxwt\"/\"default-dockercfg-v77v4\"" Apr 21 15:10:45.413419 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.413313 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-qvxwt\"/\"openshift-service-ca.crt\"" Apr 21 15:10:45.414227 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.414209 2580 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-qvxwt\"/\"kube-root-ca.crt\"" Apr 21 15:10:45.420242 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.420218 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn"] Apr 21 15:10:45.429897 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.429873 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v8h9\" (UniqueName: \"kubernetes.io/projected/8d89a3eb-c1d8-4dae-b33a-34c492592853-kube-api-access-2v8h9\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.429996 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.429914 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-proc\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.429996 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.429974 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-sys\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.430119 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.430021 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-podres\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.430180 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.430113 2580 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-lib-modules\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531476 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531444 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-lib-modules\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531672 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531495 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2v8h9\" (UniqueName: \"kubernetes.io/projected/8d89a3eb-c1d8-4dae-b33a-34c492592853-kube-api-access-2v8h9\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531672 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531516 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-proc\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531672 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531537 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-sys\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531672 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531561 2580 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-podres\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531672 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531638 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-proc\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531672 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531650 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-sys\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531672 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531648 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-lib-modules\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.531997 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.531700 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/8d89a3eb-c1d8-4dae-b33a-34c492592853-podres\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.544553 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.544528 2580 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v8h9\" (UniqueName: \"kubernetes.io/projected/8d89a3eb-c1d8-4dae-b33a-34c492592853-kube-api-access-2v8h9\") pod \"perf-node-gather-daemonset-gfpqn\" (UID: \"8d89a3eb-c1d8-4dae-b33a-34c492592853\") " pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.720528 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.720492 2580 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:45.850690 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.850661 2580 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn"] Apr 21 15:10:45.856017 ip-10-0-141-61 kubenswrapper[2580]: W0421 15:10:45.855986 2580 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8d89a3eb_c1d8_4dae_b33a_34c492592853.slice/crio-374bdf00b4d0625eec8b054e2b5bce9fe173cef9ee77b59c5753042815234cdd WatchSource:0}: Error finding container 374bdf00b4d0625eec8b054e2b5bce9fe173cef9ee77b59c5753042815234cdd: Status 404 returned error can't find the container with id 374bdf00b4d0625eec8b054e2b5bce9fe173cef9ee77b59c5753042815234cdd Apr 21 15:10:45.857769 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:45.857755 2580 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 15:10:46.390270 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:46.390226 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" event={"ID":"8d89a3eb-c1d8-4dae-b33a-34c492592853","Type":"ContainerStarted","Data":"60f76ce71634adc0196edf70c382c9c5190a14dcc974cce281fdd0c7766b7d58"} Apr 21 15:10:46.390270 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:46.390276 2580 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" event={"ID":"8d89a3eb-c1d8-4dae-b33a-34c492592853","Type":"ContainerStarted","Data":"374bdf00b4d0625eec8b054e2b5bce9fe173cef9ee77b59c5753042815234cdd"} Apr 21 15:10:46.390536 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:46.390298 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:46.409287 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:46.409231 2580 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" podStartSLOduration=1.409217129 podStartE2EDuration="1.409217129s" podCreationTimestamp="2026-04-21 15:10:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 15:10:46.406966366 +0000 UTC m=+2679.864450910" watchObservedRunningTime="2026-04-21 15:10:46.409217129 +0000 UTC m=+2679.866701635" Apr 21 15:10:46.699653 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:46.699621 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/2.log" Apr 21 15:10:46.703925 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:46.703899 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-qjbtz_1c4a98bb-ce18-4746-b5b2-979619d26ebc/console-operator/3.log" Apr 21 15:10:47.227746 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:47.227717 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-775fd649b7-h6bmr_58fc1059-03d3-4155-a810-ffd876f1e46f/console/0.log" Apr 21 15:10:47.757131 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:47.757104 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_volume-data-source-validator-7c6cbb6c87-2mnrf_898bbae2-8a55-44cc-b13f-5ab0167554e4/volume-data-source-validator/0.log" Apr 21 15:10:48.570060 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:48.570033 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-dx8mq_afea4c93-a952-4362-8775-179f3919eecc/dns/0.log" Apr 21 15:10:48.594224 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:48.594180 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-dx8mq_afea4c93-a952-4362-8775-179f3919eecc/kube-rbac-proxy/0.log" Apr 21 15:10:48.769063 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:48.769030 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-lmrrn_ccae8641-954f-48b3-85eb-812164adcec6/dns-node-resolver/0.log" Apr 21 15:10:49.282479 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:49.282449 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_image-registry-79f755579b-8wxj4_0d04ceb9-3ece-4a49-997b-3c8a3e6fc1f5/registry/0.log" Apr 21 15:10:49.358504 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:49.358466 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-89ch8_9ed791bd-825b-4608-ad66-0304c5a34a75/node-ca/0.log" Apr 21 15:10:50.396229 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:50.396133 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_data-science-gateway-data-science-gateway-class-5cb8b776cfj5p9d_37d51f19-e7f6-4871-8cb3-289134a6eabb/istio-proxy/0.log" Apr 21 15:10:50.745424 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:50.745383 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_maas-default-gateway-openshift-default-58b6f876-vbvzt_9017a28a-4013-4114-b8fb-4d3612657389/istio-proxy/0.log" Apr 21 15:10:50.772332 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:50.772309 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-6d6c54d4dd-mb7h8_0284ee69-97a8-4c33-98ee-1bb75e7f461b/router/0.log" Apr 21 15:10:51.374271 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:51.374230 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-tfksq_5badb5d3-1012-4fbf-81cc-b5190659075e/serve-healthcheck-canary/0.log" Apr 21 15:10:51.834310 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:51.834274 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-ckmd7_4c20781a-bb09-4f47-8dcd-d4d713caac69/insights-operator/0.log" Apr 21 15:10:51.834933 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:51.834753 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-ckmd7_4c20781a-bb09-4f47-8dcd-d4d713caac69/insights-operator/1.log" Apr 21 15:10:51.991913 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:51.991886 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-ncz6v_95687f91-0372-4f50-b65d-1c77acceeaeb/kube-rbac-proxy/0.log" Apr 21 15:10:52.014585 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:52.014528 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-ncz6v_95687f91-0372-4f50-b65d-1c77acceeaeb/exporter/0.log" Apr 21 15:10:52.037425 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:52.037399 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-ncz6v_95687f91-0372-4f50-b65d-1c77acceeaeb/extractor/0.log" Apr 21 15:10:52.404311 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:52.404280 2580 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-qvxwt/perf-node-gather-daemonset-gfpqn" Apr 21 15:10:54.190810 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:54.190775 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_maas-api-5c64fc8cbc-w9f4j_ed7f3c77-cdc2-4f59-939c-1bb0703ee060/maas-api/0.log" Apr 21 15:10:54.249354 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:54.249324 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_maas-controller-744d89cb86-bjwjg_b7ed0f19-f795-4675-a1d4-890a5a77d842/manager/0.log" Apr 21 15:10:54.338028 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:54.337996 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_opendatahub-operator-controller-manager-7df645bd74-9mxmk_fbebfd28-f8f1-40d8-83f8-99bbbb82a630/manager/0.log" Apr 21 15:10:56.109015 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:10:56.108975 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-lws-operator_lws-controller-manager-55dd4758fb-fkh4w_f445aa9d-90a6-44fa-9045-078ddba5fc21/manager/0.log" Apr 21 15:11:01.069030 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:01.068999 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-2lrpd_07d5eb46-986b-4f0e-9f2b-6db3b006c975/kube-storage-version-migrator-operator/1.log" Apr 21 15:11:01.069893 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:01.069866 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-2lrpd_07d5eb46-986b-4f0e-9f2b-6db3b006c975/kube-storage-version-migrator-operator/0.log" Apr 21 15:11:02.100938 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.100910 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rftw_ef45360a-730f-4594-a33e-4ae5f5cd4757/kube-multus/0.log" Apr 21 15:11:02.563442 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.563404 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-sz9sg_29a15288-a939-4c42-b167-118380ab5462/kube-multus-additional-cni-plugins/0.log" Apr 21 15:11:02.588976 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.588948 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-sz9sg_29a15288-a939-4c42-b167-118380ab5462/egress-router-binary-copy/0.log" Apr 21 15:11:02.611879 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.611852 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-sz9sg_29a15288-a939-4c42-b167-118380ab5462/cni-plugins/0.log" Apr 21 15:11:02.634371 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.634345 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-sz9sg_29a15288-a939-4c42-b167-118380ab5462/bond-cni-plugin/0.log" Apr 21 15:11:02.658505 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.658481 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-sz9sg_29a15288-a939-4c42-b167-118380ab5462/routeoverride-cni/0.log" Apr 21 15:11:02.682632 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.682598 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-sz9sg_29a15288-a939-4c42-b167-118380ab5462/whereabouts-cni-bincopy/0.log" Apr 21 15:11:02.711330 ip-10-0-141-61 kubenswrapper[2580]: I0421 15:11:02.711303 2580 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-sz9sg_29a15288-a939-4c42-b167-118380ab5462/whereabouts-cni/0.log"