Apr 22 18:17:57.947678 ip-10-0-141-172 systemd[1]: kubelet.service: Failed to load environment files: No such file or directory Apr 22 18:17:57.947692 ip-10-0-141-172 systemd[1]: kubelet.service: Failed to run 'start-pre' task: No such file or directory Apr 22 18:17:57.947701 ip-10-0-141-172 systemd[1]: kubelet.service: Failed with result 'resources'. Apr 22 18:17:57.948012 ip-10-0-141-172 systemd[1]: Failed to start Kubernetes Kubelet. Apr 22 18:18:08.069847 ip-10-0-141-172 systemd[1]: kubelet.service: Failed to schedule restart job: Unit crio.service not found. Apr 22 18:18:08.069861 ip-10-0-141-172 systemd[1]: kubelet.service: Failed with result 'resources'. -- Boot 0a22e9480ec341579c7eae7cf21b81f1 -- Apr 22 18:20:43.078241 ip-10-0-141-172 systemd[1]: Starting Kubernetes Kubelet... Apr 22 18:20:43.505597 ip-10-0-141-172 kubenswrapper[2581]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 18:20:43.505597 ip-10-0-141-172 kubenswrapper[2581]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 22 18:20:43.505597 ip-10-0-141-172 kubenswrapper[2581]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 18:20:43.505597 ip-10-0-141-172 kubenswrapper[2581]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 22 18:20:43.505597 ip-10-0-141-172 kubenswrapper[2581]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 18:20:43.506781 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.506691 2581 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 22 18:20:43.509707 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509686 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 18:20:43.509707 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509703 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 18:20:43.509707 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509707 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 18:20:43.509707 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509711 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 18:20:43.509707 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509715 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509719 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509722 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509727 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509730 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509734 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509737 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509741 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509745 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509748 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509751 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509756 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509759 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509762 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509766 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509769 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509773 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509777 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509780 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509784 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 18:20:43.510011 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509788 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509791 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509795 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509798 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509803 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509806 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509810 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509814 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509818 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509822 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509826 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509830 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509834 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509839 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509843 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509848 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509852 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509856 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509860 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 18:20:43.510877 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509864 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509868 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509871 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509876 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509880 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509884 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509888 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509895 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509901 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509905 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509909 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509914 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509935 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509941 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509945 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509950 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509954 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509959 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509963 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509967 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 18:20:43.511691 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509971 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509975 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509978 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509982 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509986 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509990 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509994 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.509999 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510005 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510008 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510012 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510016 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510021 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510027 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510033 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510037 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510041 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510046 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510050 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 18:20:43.512525 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510054 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510058 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510062 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510065 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510629 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510638 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510642 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510646 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510650 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510654 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510658 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510662 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510666 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510670 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510674 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510678 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510682 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510685 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510690 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 18:20:43.512991 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510695 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510703 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510710 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510715 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510720 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510724 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510728 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510732 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510736 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510740 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510744 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510747 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510752 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510755 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510759 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510763 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510767 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510771 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510775 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 18:20:43.513499 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510780 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510784 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510788 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510793 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510798 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510801 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510805 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510809 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510813 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510817 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510821 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510825 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510829 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510832 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510837 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510841 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510846 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510850 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510854 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510858 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 18:20:43.514061 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510862 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510866 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510870 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510874 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510878 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510881 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510886 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510900 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510904 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510907 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510911 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510935 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510941 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510945 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510949 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510953 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510957 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510961 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510965 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 18:20:43.514695 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510968 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510972 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510976 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510980 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510984 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510988 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.510995 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.511000 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.511004 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.511008 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.511013 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.511017 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.511021 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511121 2581 flags.go:64] FLAG: --address="0.0.0.0" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511131 2581 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511141 2581 flags.go:64] FLAG: --anonymous-auth="true" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511147 2581 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511153 2581 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511158 2581 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511164 2581 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511171 2581 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 22 18:20:43.515193 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511176 2581 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511181 2581 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511186 2581 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511192 2581 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511196 2581 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511201 2581 flags.go:64] FLAG: --cgroup-root="" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511205 2581 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511210 2581 flags.go:64] FLAG: --client-ca-file="" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511214 2581 flags.go:64] FLAG: --cloud-config="" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511218 2581 flags.go:64] FLAG: --cloud-provider="external" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511223 2581 flags.go:64] FLAG: --cluster-dns="[]" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511228 2581 flags.go:64] FLAG: --cluster-domain="" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511232 2581 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511237 2581 flags.go:64] FLAG: --config-dir="" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511241 2581 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511247 2581 flags.go:64] FLAG: --container-log-max-files="5" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511253 2581 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511257 2581 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511263 2581 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511268 2581 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511272 2581 flags.go:64] FLAG: --contention-profiling="false" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511277 2581 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511282 2581 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511287 2581 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511292 2581 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 22 18:20:43.515767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511298 2581 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511303 2581 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511307 2581 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511312 2581 flags.go:64] FLAG: --enable-load-reader="false" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511316 2581 flags.go:64] FLAG: --enable-server="true" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511321 2581 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511327 2581 flags.go:64] FLAG: --event-burst="100" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511331 2581 flags.go:64] FLAG: --event-qps="50" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511336 2581 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511340 2581 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511345 2581 flags.go:64] FLAG: --eviction-hard="" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511352 2581 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511356 2581 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511361 2581 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511366 2581 flags.go:64] FLAG: --eviction-soft="" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511370 2581 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511375 2581 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511383 2581 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511388 2581 flags.go:64] FLAG: --experimental-mounter-path="" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511392 2581 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511396 2581 flags.go:64] FLAG: --fail-swap-on="true" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511401 2581 flags.go:64] FLAG: --feature-gates="" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511406 2581 flags.go:64] FLAG: --file-check-frequency="20s" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511411 2581 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511416 2581 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 22 18:20:43.516402 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511421 2581 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511426 2581 flags.go:64] FLAG: --healthz-port="10248" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511431 2581 flags.go:64] FLAG: --help="false" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511435 2581 flags.go:64] FLAG: --hostname-override="ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511440 2581 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511445 2581 flags.go:64] FLAG: --http-check-frequency="20s" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511450 2581 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511455 2581 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511460 2581 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511465 2581 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511470 2581 flags.go:64] FLAG: --image-service-endpoint="" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511475 2581 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511479 2581 flags.go:64] FLAG: --kube-api-burst="100" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511484 2581 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511489 2581 flags.go:64] FLAG: --kube-api-qps="50" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511493 2581 flags.go:64] FLAG: --kube-reserved="" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511498 2581 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511502 2581 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511507 2581 flags.go:64] FLAG: --kubelet-cgroups="" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511511 2581 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511515 2581 flags.go:64] FLAG: --lock-file="" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511519 2581 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511524 2581 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511529 2581 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 22 18:20:43.517010 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511539 2581 flags.go:64] FLAG: --log-json-split-stream="false" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511544 2581 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511567 2581 flags.go:64] FLAG: --log-text-split-stream="false" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511573 2581 flags.go:64] FLAG: --logging-format="text" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511578 2581 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511583 2581 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511587 2581 flags.go:64] FLAG: --manifest-url="" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511592 2581 flags.go:64] FLAG: --manifest-url-header="" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511598 2581 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511603 2581 flags.go:64] FLAG: --max-open-files="1000000" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511609 2581 flags.go:64] FLAG: --max-pods="110" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511613 2581 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511618 2581 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511622 2581 flags.go:64] FLAG: --memory-manager-policy="None" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511628 2581 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511632 2581 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511637 2581 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511647 2581 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511658 2581 flags.go:64] FLAG: --node-status-max-images="50" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511663 2581 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511668 2581 flags.go:64] FLAG: --oom-score-adj="-999" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511673 2581 flags.go:64] FLAG: --pod-cidr="" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511677 2581 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 22 18:20:43.517581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511686 2581 flags.go:64] FLAG: --pod-manifest-path="" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511690 2581 flags.go:64] FLAG: --pod-max-pids="-1" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511695 2581 flags.go:64] FLAG: --pods-per-core="0" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511699 2581 flags.go:64] FLAG: --port="10250" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511704 2581 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511708 2581 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-095f55e1d0165d90f" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511713 2581 flags.go:64] FLAG: --qos-reserved="" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511717 2581 flags.go:64] FLAG: --read-only-port="10255" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511722 2581 flags.go:64] FLAG: --register-node="true" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511727 2581 flags.go:64] FLAG: --register-schedulable="true" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511733 2581 flags.go:64] FLAG: --register-with-taints="" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511744 2581 flags.go:64] FLAG: --registry-burst="10" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511749 2581 flags.go:64] FLAG: --registry-qps="5" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511753 2581 flags.go:64] FLAG: --reserved-cpus="" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511758 2581 flags.go:64] FLAG: --reserved-memory="" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511764 2581 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511769 2581 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511774 2581 flags.go:64] FLAG: --rotate-certificates="false" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511778 2581 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511783 2581 flags.go:64] FLAG: --runonce="false" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511787 2581 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511792 2581 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511796 2581 flags.go:64] FLAG: --seccomp-default="false" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511801 2581 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511806 2581 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511811 2581 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 22 18:20:43.518132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511817 2581 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511822 2581 flags.go:64] FLAG: --storage-driver-password="root" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511826 2581 flags.go:64] FLAG: --storage-driver-secure="false" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511831 2581 flags.go:64] FLAG: --storage-driver-table="stats" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511835 2581 flags.go:64] FLAG: --storage-driver-user="root" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511839 2581 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511844 2581 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511849 2581 flags.go:64] FLAG: --system-cgroups="" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511853 2581 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511861 2581 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511866 2581 flags.go:64] FLAG: --tls-cert-file="" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511870 2581 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511875 2581 flags.go:64] FLAG: --tls-min-version="" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511879 2581 flags.go:64] FLAG: --tls-private-key-file="" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511884 2581 flags.go:64] FLAG: --topology-manager-policy="none" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511888 2581 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511894 2581 flags.go:64] FLAG: --topology-manager-scope="container" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511899 2581 flags.go:64] FLAG: --v="2" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511905 2581 flags.go:64] FLAG: --version="false" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511911 2581 flags.go:64] FLAG: --vmodule="" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511938 2581 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.511944 2581 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512085 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512091 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 18:20:43.519023 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512096 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512100 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512105 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512110 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512114 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512118 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512123 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512134 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512140 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512144 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512148 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512152 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512156 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512161 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512165 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512169 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512175 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512181 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512185 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 18:20:43.519876 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512190 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512195 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512199 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512202 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512207 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512212 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512216 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512220 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512224 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512228 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512232 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512236 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512240 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512244 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512248 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512252 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512256 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512260 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512264 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512271 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 18:20:43.520422 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512276 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512283 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512288 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512292 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512296 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512300 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512304 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512308 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512311 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512315 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512319 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512323 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512327 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512331 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512335 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512339 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512342 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512347 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512351 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 18:20:43.520891 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512355 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512360 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512364 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512368 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512372 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512376 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512380 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512383 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512387 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512391 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512395 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512399 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512403 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512407 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512412 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512417 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512421 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512425 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512429 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512433 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 18:20:43.521354 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512437 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 18:20:43.521831 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512441 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 18:20:43.521831 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512444 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 18:20:43.521831 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512448 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 18:20:43.521831 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512452 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 18:20:43.521831 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.512456 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 18:20:43.521831 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.513390 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 18:20:43.522024 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.522003 2581 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 22 18:20:43.522055 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.522025 2581 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 22 18:20:43.522084 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522068 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 18:20:43.522084 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522073 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 18:20:43.522084 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522076 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 18:20:43.522084 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522079 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 18:20:43.522084 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522082 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 18:20:43.522084 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522085 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522088 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522091 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522094 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522096 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522099 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522102 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522104 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522107 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522110 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522112 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522115 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522117 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522120 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522127 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522130 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522133 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522135 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522138 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522140 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 18:20:43.522227 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522142 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522145 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522147 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522150 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522153 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522156 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522160 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522163 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522166 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522168 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522171 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522173 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522176 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522178 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522181 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522184 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522186 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522188 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522191 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522193 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 18:20:43.522733 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522195 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522198 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522200 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522202 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522205 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522207 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522209 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522212 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522215 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522217 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522220 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522222 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522225 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522227 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522229 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522232 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522234 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522237 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522240 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 18:20:43.523312 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522242 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522244 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522247 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522249 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522251 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522253 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522256 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522259 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522261 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522264 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522266 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522269 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522271 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522274 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522276 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522280 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522284 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522287 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522290 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522292 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 18:20:43.523758 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522295 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522297 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.522302 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522396 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522400 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522403 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522406 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522409 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522411 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522414 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522417 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522419 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522422 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522425 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522427 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522430 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 18:20:43.524239 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522432 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522434 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522437 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522439 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522442 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522445 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522447 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522450 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522454 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522457 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522460 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522462 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522465 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522467 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522470 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522472 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522475 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522477 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522480 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522482 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 18:20:43.524637 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522484 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522487 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522489 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522491 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522494 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522496 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522498 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522501 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522503 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522506 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522508 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522510 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522512 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522516 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522519 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522522 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522525 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522527 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522530 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 18:20:43.525114 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522533 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522535 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522537 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522540 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522543 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522545 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522548 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522550 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522552 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522555 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522558 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522560 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522563 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522565 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522567 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522569 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522572 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522574 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522576 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522578 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 18:20:43.525561 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522581 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522583 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522586 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522588 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522590 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522592 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522595 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522597 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522599 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522602 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522604 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522607 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522609 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:43.522612 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.522616 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 18:20:43.526039 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.522712 2581 server.go:962] "Client rotation is on, will bootstrap in background" Apr 22 18:20:43.526427 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.526414 2581 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 22 18:20:43.527278 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.527266 2581 server.go:1019] "Starting client certificate rotation" Apr 22 18:20:43.527382 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.527365 2581 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 18:20:43.527436 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.527415 2581 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 18:20:43.550865 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.550845 2581 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 18:20:43.552454 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.552433 2581 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 18:20:43.565479 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.565461 2581 log.go:25] "Validated CRI v1 runtime API" Apr 22 18:20:43.570782 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.570768 2581 log.go:25] "Validated CRI v1 image API" Apr 22 18:20:43.572955 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.572938 2581 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 22 18:20:43.577160 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.577143 2581 fs.go:135] Filesystem UUIDs: map[10f77c16-59cb-4bfc-8a4f-ad3b6b046600:/dev/nvme0n1p4 7B77-95E7:/dev/nvme0n1p2 a138b2ec-c9a9-4efc-9a42-5252a0791a49:/dev/nvme0n1p3] Apr 22 18:20:43.577209 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.577161 2581 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 22 18:20:43.582334 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.582231 2581 manager.go:217] Machine: {Timestamp:2026-04-22 18:20:43.581058158 +0000 UTC m=+0.385121514 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3162009 MemoryCapacity:32812163072 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec20acec1a01b45683fa250d73e68f84 SystemUUID:ec20acec-1a01-b456-83fa-250d73e68f84 BootID:0a22e948-0ec3-4157-9c7e-ae7cf21b81f1 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16406081536 Type:vfs Inodes:4005391 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6562435072 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16406081536 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:82:f5:73:26:f1 Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:82:f5:73:26:f1 Speed:0 Mtu:9001} {Name:ovs-system MacAddress:66:bf:ed:75:15:7c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:32812163072 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:34603008 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 22 18:20:43.582334 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.582324 2581 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 22 18:20:43.582447 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.582390 2581 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 22 18:20:43.583511 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.583475 2581 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 22 18:20:43.583632 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.583512 2581 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-141-172.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 22 18:20:43.583689 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.583641 2581 topology_manager.go:138] "Creating topology manager with none policy" Apr 22 18:20:43.583689 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.583650 2581 container_manager_linux.go:306] "Creating device plugin manager" Apr 22 18:20:43.583689 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.583663 2581 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 18:20:43.583689 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.583681 2581 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 18:20:43.585354 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.585343 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 22 18:20:43.585562 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.585535 2581 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 18:20:43.585644 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.585612 2581 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 22 18:20:43.588232 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.588222 2581 kubelet.go:491] "Attempting to sync node with API server" Apr 22 18:20:43.588269 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.588235 2581 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 22 18:20:43.588876 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.588868 2581 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 22 18:20:43.588913 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.588880 2581 kubelet.go:397] "Adding apiserver pod source" Apr 22 18:20:43.588913 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.588889 2581 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 22 18:20:43.589901 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.589890 2581 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 18:20:43.589959 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.589908 2581 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 18:20:43.592528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.592513 2581 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 22 18:20:43.594199 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.594184 2581 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 22 18:20:43.595432 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595418 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595439 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595447 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595456 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595464 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595472 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595481 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595489 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595498 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595507 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 22 18:20:43.595517 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595519 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 22 18:20:43.595833 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.595532 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 22 18:20:43.596374 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.596362 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 22 18:20:43.596421 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.596376 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 22 18:20:43.599659 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.599642 2581 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-141-172.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 22 18:20:43.599988 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.599976 2581 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 22 18:20:43.600062 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.600016 2581 server.go:1295] "Started kubelet" Apr 22 18:20:43.600125 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.600085 2581 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 22 18:20:43.600125 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.600114 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-141-172.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 22 18:20:43.600255 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.600153 2581 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 22 18:20:43.600255 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.600196 2581 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 22 18:20:43.600747 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.600120 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 22 18:20:43.601058 ip-10-0-141-172 systemd[1]: Started Kubernetes Kubelet. Apr 22 18:20:43.602348 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.602331 2581 server.go:317] "Adding debug handlers to kubelet server" Apr 22 18:20:43.603878 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.603861 2581 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 22 18:20:43.608273 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.608257 2581 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 22 18:20:43.609057 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.609042 2581 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 22 18:20:43.609259 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.608171 2581 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-141-172.ec2.internal.18a8c0cde88c2526 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-141-172.ec2.internal,UID:ip-10-0-141-172.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-10-0-141-172.ec2.internal,},FirstTimestamp:2026-04-22 18:20:43.599988006 +0000 UTC m=+0.404051365,LastTimestamp:2026-04-22 18:20:43.599988006 +0000 UTC m=+0.404051365,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-141-172.ec2.internal,}" Apr 22 18:20:43.609656 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.609636 2581 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 22 18:20:43.609723 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.609660 2581 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 22 18:20:43.609777 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.609732 2581 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 22 18:20:43.609895 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.609856 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:43.609967 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.609899 2581 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 22 18:20:43.610125 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610109 2581 factory.go:55] Registering systemd factory Apr 22 18:20:43.610189 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610165 2581 factory.go:223] Registration of the systemd container factory successfully Apr 22 18:20:43.610239 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610110 2581 reconstruct.go:97] "Volume reconstruction finished" Apr 22 18:20:43.610239 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610211 2581 reconciler.go:26] "Reconciler: start to sync state" Apr 22 18:20:43.610364 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610352 2581 factory.go:153] Registering CRI-O factory Apr 22 18:20:43.610364 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610365 2581 factory.go:223] Registration of the crio container factory successfully Apr 22 18:20:43.610438 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610409 2581 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 22 18:20:43.610438 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610426 2581 factory.go:103] Registering Raw factory Apr 22 18:20:43.610438 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610435 2581 manager.go:1196] Started watching for new ooms in manager Apr 22 18:20:43.610867 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.610846 2581 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-dcbp8" Apr 22 18:20:43.611860 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.611830 2581 manager.go:319] Starting recovery of all containers Apr 22 18:20:43.614449 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.614281 2581 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"ip-10-0-141-172.ec2.internal\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Apr 22 18:20:43.614449 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.614374 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 22 18:20:43.619222 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.619202 2581 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-dcbp8" Apr 22 18:20:43.624125 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.624017 2581 manager.go:324] Recovery completed Apr 22 18:20:43.627858 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.627846 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 18:20:43.631561 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.631547 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientMemory" Apr 22 18:20:43.631615 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.631572 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 18:20:43.631615 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.631582 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientPID" Apr 22 18:20:43.632034 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.632019 2581 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 22 18:20:43.632034 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.632033 2581 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 22 18:20:43.632176 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.632049 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 22 18:20:43.633368 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.633277 2581 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-141-172.ec2.internal.18a8c0cdea6de6f6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-141-172.ec2.internal,UID:ip-10-0-141-172.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node ip-10-0-141-172.ec2.internal status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:ip-10-0-141-172.ec2.internal,},FirstTimestamp:2026-04-22 18:20:43.631560438 +0000 UTC m=+0.435623794,LastTimestamp:2026-04-22 18:20:43.631560438 +0000 UTC m=+0.435623794,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-141-172.ec2.internal,}" Apr 22 18:20:43.635561 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.635547 2581 policy_none.go:49] "None policy: Start" Apr 22 18:20:43.635636 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.635565 2581 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 22 18:20:43.635636 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.635577 2581 state_mem.go:35] "Initializing new in-memory state store" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.677396 2581 manager.go:341] "Starting Device Plugin manager" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.677438 2581 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.677450 2581 server.go:85] "Starting device plugin registration server" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.677657 2581 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.677668 2581 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.677754 2581 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.677820 2581 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.677829 2581 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.678424 2581 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 22 18:20:43.687351 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.678459 2581 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:43.736911 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.736876 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 22 18:20:43.738202 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.738183 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 22 18:20:43.738665 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.738652 2581 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 22 18:20:43.738720 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.738682 2581 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 22 18:20:43.738720 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.738689 2581 kubelet.go:2451] "Starting kubelet main sync loop" Apr 22 18:20:43.738789 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.738722 2581 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 22 18:20:43.741547 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.741528 2581 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 18:20:43.777829 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.777786 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 18:20:43.778523 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.778508 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientMemory" Apr 22 18:20:43.778573 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.778534 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 18:20:43.778573 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.778544 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientPID" Apr 22 18:20:43.778573 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.778564 2581 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.786854 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.786842 2581 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.786904 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.786860 2581 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-141-172.ec2.internal\": node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:43.800457 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.800437 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:43.839047 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.839025 2581 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal"] Apr 22 18:20:43.839126 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.839104 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 18:20:43.840609 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.840596 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientMemory" Apr 22 18:20:43.840676 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.840620 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 18:20:43.840676 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.840630 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientPID" Apr 22 18:20:43.842612 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.842600 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 18:20:43.842746 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.842732 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.842799 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.842765 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 18:20:43.845448 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.845432 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientMemory" Apr 22 18:20:43.845528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.845432 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientMemory" Apr 22 18:20:43.845528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.845481 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 18:20:43.845528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.845491 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientPID" Apr 22 18:20:43.845623 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.845464 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 18:20:43.845623 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.845558 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientPID" Apr 22 18:20:43.847451 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.847436 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.847522 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.847457 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 18:20:43.848092 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.848068 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientMemory" Apr 22 18:20:43.848092 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.848089 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 18:20:43.848216 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.848098 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeHasSufficientPID" Apr 22 18:20:43.869390 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.869372 2581 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-141-172.ec2.internal\" not found" node="ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.873764 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.873749 2581 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-141-172.ec2.internal\" not found" node="ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.901275 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:43.901252 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:43.911690 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.911669 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/060906b545f652629bb1b4f63cd4a0c8-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal\" (UID: \"060906b545f652629bb1b4f63cd4a0c8\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.911747 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.911700 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/060906b545f652629bb1b4f63cd4a0c8-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal\" (UID: \"060906b545f652629bb1b4f63cd4a0c8\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:43.911747 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:43.911717 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/b128912d2b3d8968717857155603d30b-config\") pod \"kube-apiserver-proxy-ip-10-0-141-172.ec2.internal\" (UID: \"b128912d2b3d8968717857155603d30b\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.001531 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.001506 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.012436 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.012418 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/060906b545f652629bb1b4f63cd4a0c8-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal\" (UID: \"060906b545f652629bb1b4f63cd4a0c8\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.012512 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.012448 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/060906b545f652629bb1b4f63cd4a0c8-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal\" (UID: \"060906b545f652629bb1b4f63cd4a0c8\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.012512 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.012471 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/b128912d2b3d8968717857155603d30b-config\") pod \"kube-apiserver-proxy-ip-10-0-141-172.ec2.internal\" (UID: \"b128912d2b3d8968717857155603d30b\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.012512 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.012504 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/b128912d2b3d8968717857155603d30b-config\") pod \"kube-apiserver-proxy-ip-10-0-141-172.ec2.internal\" (UID: \"b128912d2b3d8968717857155603d30b\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.012629 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.012516 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/060906b545f652629bb1b4f63cd4a0c8-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal\" (UID: \"060906b545f652629bb1b4f63cd4a0c8\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.012629 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.012506 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/060906b545f652629bb1b4f63cd4a0c8-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal\" (UID: \"060906b545f652629bb1b4f63cd4a0c8\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.101855 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.101805 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.171245 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.171227 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.175536 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.175521 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" Apr 22 18:20:44.202044 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.202014 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.302423 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.302403 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.402902 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.402844 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.503423 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.503402 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.526836 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.526821 2581 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 22 18:20:44.527298 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.526975 2581 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 22 18:20:44.604388 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.604369 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.608465 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.608449 2581 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 22 18:20:44.620751 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.620728 2581 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 18:20:44.620842 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.620734 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-21 18:15:43 +0000 UTC" deadline="2027-12-27 15:10:21.31702376 +0000 UTC" Apr 22 18:20:44.620842 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.620765 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="14732h49m36.696260961s" Apr 22 18:20:44.642061 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.642041 2581 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-w5kvk" Apr 22 18:20:44.649898 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.649879 2581 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-w5kvk" Apr 22 18:20:44.658654 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.658603 2581 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 18:20:44.705145 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.705109 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.726371 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:44.726322 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod060906b545f652629bb1b4f63cd4a0c8.slice/crio-0f9518a97def2e7b726e55bd650cd17eac676998b07ad3ba07bf5546746f0eca WatchSource:0}: Error finding container 0f9518a97def2e7b726e55bd650cd17eac676998b07ad3ba07bf5546746f0eca: Status 404 returned error can't find the container with id 0f9518a97def2e7b726e55bd650cd17eac676998b07ad3ba07bf5546746f0eca Apr 22 18:20:44.726648 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:44.726624 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb128912d2b3d8968717857155603d30b.slice/crio-ea2e0253a590fb2138b8d8b5c11df41d8edc94bbbfe354f3f7f2850961116e4b WatchSource:0}: Error finding container ea2e0253a590fb2138b8d8b5c11df41d8edc94bbbfe354f3f7f2850961116e4b: Status 404 returned error can't find the container with id ea2e0253a590fb2138b8d8b5c11df41d8edc94bbbfe354f3f7f2850961116e4b Apr 22 18:20:44.730647 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.730632 2581 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 18:20:44.741712 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.741669 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" event={"ID":"b128912d2b3d8968717857155603d30b","Type":"ContainerStarted","Data":"ea2e0253a590fb2138b8d8b5c11df41d8edc94bbbfe354f3f7f2850961116e4b"} Apr 22 18:20:44.742541 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.742523 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" event={"ID":"060906b545f652629bb1b4f63cd4a0c8","Type":"ContainerStarted","Data":"0f9518a97def2e7b726e55bd650cd17eac676998b07ad3ba07bf5546746f0eca"} Apr 22 18:20:44.804584 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:44.804566 2581 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 18:20:44.805556 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.805535 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:44.906303 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:44.906279 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:45.006799 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.006740 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-141-172.ec2.internal\" not found" Apr 22 18:20:45.098408 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.098259 2581 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 18:20:45.109894 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.109649 2581 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" Apr 22 18:20:45.122223 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.122121 2581 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 18:20:45.123131 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.123072 2581 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" Apr 22 18:20:45.133300 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.133279 2581 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 18:20:45.589388 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.589365 2581 apiserver.go:52] "Watching apiserver" Apr 22 18:20:45.597387 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.597364 2581 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 22 18:20:45.597739 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.597720 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-pqljk","openshift-image-registry/node-ca-qxft5","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal","openshift-multus/multus-additional-cni-plugins-x2fkn","openshift-multus/multus-p27q9","openshift-multus/network-metrics-daemon-xl5pv","openshift-network-diagnostics/network-check-target-47xwd","openshift-network-operator/iptables-alerter-96txh","openshift-ovn-kubernetes/ovnkube-node-44fqr","kube-system/konnectivity-agent-v2tgg","kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45","openshift-cluster-node-tuning-operator/tuned-zhjmm"] Apr 22 18:20:45.602903 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.602869 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:45.603009 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.602973 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:45.605136 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.605117 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.605221 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.605203 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.607410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.607390 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.608281 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.608259 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 22 18:20:45.608523 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.608505 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 22 18:20:45.608606 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.608536 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-c4q6z\"" Apr 22 18:20:45.608735 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.608720 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-xbbvw\"" Apr 22 18:20:45.608905 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.608892 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 22 18:20:45.608983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.608965 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 22 18:20:45.609077 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.609063 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 22 18:20:45.609114 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.608898 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 22 18:20:45.609237 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.609224 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 22 18:20:45.609301 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.609289 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 22 18:20:45.609628 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.609610 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:45.609721 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.609695 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:20:45.609783 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.609755 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-dpjth\"" Apr 22 18:20:45.610316 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.610298 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 22 18:20:45.612576 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.612552 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.614828 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.614811 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.615134 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.614984 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 22 18:20:45.615134 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.615013 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-q64zh\"" Apr 22 18:20:45.615286 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.615268 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 22 18:20:45.621350 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.621328 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 22 18:20:45.621525 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.621508 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 22 18:20:45.621811 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.621797 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622075 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-tmp-dir\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622118 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-serviceca\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622151 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5224k\" (UniqueName: \"kubernetes.io/projected/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-kube-api-access-5224k\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622182 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-socket-dir-parent\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622212 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-k8s-cni-cncf-io\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622250 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-cni-multus\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622275 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-cnibin\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622310 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-hostroot\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622337 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-etc-kubernetes\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622356 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622374 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-os-release\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622390 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-tuning-conf-dir\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622407 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5m6l\" (UniqueName: \"kubernetes.io/projected/c0602c54-63d4-4509-bf26-1bfa53e55b60-kube-api-access-s5m6l\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622425 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-os-release\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622448 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-cni-binary-copy\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.623832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622475 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-hosts-file\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622497 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-system-cni-dir\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622526 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622564 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-cnibin\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622583 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-netns\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622600 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-multus-certs\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622619 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622636 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-cni-bin\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622657 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-daemon-config\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622675 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88b8b\" (UniqueName: \"kubernetes.io/projected/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-kube-api-access-88b8b\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622692 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-jts9f\"" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622705 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnkf8\" (UniqueName: \"kubernetes.io/projected/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-kube-api-access-gnkf8\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622734 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-cni-binary-copy\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622765 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622792 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-system-cni-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622820 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-conf-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622860 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw47q\" (UniqueName: \"kubernetes.io/projected/b5ac120e-7d32-45b8-a5d6-57aa386e732f-kube-api-access-mw47q\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:45.624610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622896 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-host\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.625322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622973 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-cni-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.625322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.623005 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-kubelet\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.625322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.623164 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.625322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.623203 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.625322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.622601 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 22 18:20:45.626049 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.625823 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.627623 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.627607 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 22 18:20:45.629321 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.629303 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 22 18:20:45.629556 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.629540 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 22 18:20:45.629660 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.629643 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-l4w5z\"" Apr 22 18:20:45.630663 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.630615 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 22 18:20:45.630745 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.630730 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 22 18:20:45.630826 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.630782 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-2w9sz\"" Apr 22 18:20:45.630961 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.630944 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 22 18:20:45.635459 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.635441 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 22 18:20:45.635714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.635686 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 22 18:20:45.635819 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.635798 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 22 18:20:45.635903 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.635877 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 22 18:20:45.636078 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.636062 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 22 18:20:45.636902 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.636751 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 22 18:20:45.636902 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.636766 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-kqgx2\"" Apr 22 18:20:45.636902 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.636793 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 22 18:20:45.636902 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.636855 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-245xn\"" Apr 22 18:20:45.650699 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.650672 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 18:15:44 +0000 UTC" deadline="2027-12-06 00:36:49.657370932 +0000 UTC" Apr 22 18:20:45.650798 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.650699 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14214h16m4.006674737s" Apr 22 18:20:45.693214 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.693190 2581 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 18:20:45.710742 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.710724 2581 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 22 18:20:45.723894 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.723874 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-os-release\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724011 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.723906 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-cni-binary-copy\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724011 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.723952 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0ad3985-82e5-4fed-89ff-cd6f30087d98-host-slash\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.724011 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.723978 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-cnibin\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724011 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724002 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-multus-certs\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724191 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724003 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-os-release\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724191 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724029 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:45.724191 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724057 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-run-netns\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.724191 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724066 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-multus-certs\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724191 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724076 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-cnibin\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724191 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724080 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-etc-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.724191 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.724189 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724217 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-modprobe-d\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.724241 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs podName:b5ac120e-7d32-45b8-a5d6-57aa386e732f nodeName:}" failed. No retries permitted until 2026-04-22 18:20:46.224226006 +0000 UTC m=+3.028289349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs") pod "network-metrics-daemon-xl5pv" (UID: "b5ac120e-7d32-45b8-a5d6-57aa386e732f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724272 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysconfig\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724292 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-cni-bin\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724311 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-daemon-config\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724345 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-88b8b\" (UniqueName: \"kubernetes.io/projected/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-kube-api-access-88b8b\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724348 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-cni-bin\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724360 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-slash\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724386 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724430 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-sys\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.724462 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724468 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gnkf8\" (UniqueName: \"kubernetes.io/projected/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-kube-api-access-gnkf8\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724495 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724523 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-node-log\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724548 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/c32b819b-1ae9-498a-876b-87995a803b2c-agent-certs\") pod \"konnectivity-agent-v2tgg\" (UID: \"c32b819b-1ae9-498a-876b-87995a803b2c\") " pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724550 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-cni-binary-copy\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724573 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-cni-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724616 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-kubelet\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724641 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/a0ad3985-82e5-4fed-89ff-cd6f30087d98-iptables-alerter-script\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724665 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-systemd-units\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724688 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-cni-netd\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724700 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-kubelet\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724712 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovnkube-script-lib\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724735 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdt2q\" (UniqueName: \"kubernetes.io/projected/938a2b8c-12e5-458f-88a8-d8c54b3148e0-kube-api-access-vdt2q\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724663 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-cni-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724774 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/721979eb-ddad-4268-a95a-833cf12d6074-etc-tuned\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724811 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-serviceca\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724834 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5224k\" (UniqueName: \"kubernetes.io/projected/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-kube-api-access-5224k\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.725009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724861 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-kubelet-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724833 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-daemon-config\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724885 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94hqc\" (UniqueName: \"kubernetes.io/projected/90936ce6-5c5f-4f38-9487-0e4631ff77e6-kube-api-access-94hqc\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724973 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-cnibin\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.724999 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-etc-kubernetes\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725033 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725045 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-var-lib-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725048 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-cnibin\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725076 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/c32b819b-1ae9-498a-876b-87995a803b2c-konnectivity-ca\") pod \"konnectivity-agent-v2tgg\" (UID: \"c32b819b-1ae9-498a-876b-87995a803b2c\") " pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725081 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-etc-kubernetes\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725113 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-lib-modules\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725139 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-var-lib-kubelet\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725182 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-ovn\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725186 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-serviceca\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725215 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725241 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-tuning-conf-dir\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725272 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-s5m6l\" (UniqueName: \"kubernetes.io/projected/c0602c54-63d4-4509-bf26-1bfa53e55b60-kube-api-access-s5m6l\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.725760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725319 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz55l\" (UniqueName: \"kubernetes.io/projected/a0ad3985-82e5-4fed-89ff-cd6f30087d98-kube-api-access-fz55l\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725345 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovnkube-config\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725371 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-tuning-conf-dir\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725377 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-hosts-file\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725437 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-hosts-file\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725419 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-system-cni-dir\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725479 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725507 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-netns\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725523 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-system-cni-dir\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725533 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-kubelet\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725560 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-systemd\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725582 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-host\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725553 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-netns\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725606 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7h2m\" (UniqueName: \"kubernetes.io/projected/721979eb-ddad-4268-a95a-833cf12d6074-kube-api-access-g7h2m\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725634 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-device-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725659 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725683 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-log-socket\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726277 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725719 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-cni-bin\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725740 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-kubernetes\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725754 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysctl-conf\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725771 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-cni-binary-copy\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725796 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-system-cni-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725844 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-system-cni-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725895 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-conf-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725942 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mw47q\" (UniqueName: \"kubernetes.io/projected/b5ac120e-7d32-45b8-a5d6-57aa386e732f-kube-api-access-mw47q\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725964 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-conf-dir\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725969 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-systemd\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.725993 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/721979eb-ddad-4268-a95a-833cf12d6074-tmp\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726017 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-host\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726017 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726057 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovn-node-metrics-cert\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726083 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-host\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726088 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-socket-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726125 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-env-overrides\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.726983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726152 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysctl-d\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726177 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-run\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726206 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-tmp-dir\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726220 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0602c54-63d4-4509-bf26-1bfa53e55b60-cni-binary-copy\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726233 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-socket-dir-parent\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726261 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-k8s-cni-cncf-io\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726288 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-cni-multus\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726314 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-etc-selinux\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726338 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-run-k8s-cni-cncf-io\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726341 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-sys-fs\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726372 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726378 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-multus-socket-dir-parent\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726400 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-hostroot\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726429 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-registration-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726444 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-hostroot\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726456 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-os-release\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726496 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0602c54-63d4-4509-bf26-1bfa53e55b60-os-release\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.727410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726780 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-tmp-dir\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.727863 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.726353 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-host-var-lib-cni-multus\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.732655 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.732637 2581 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 22 18:20:45.733940 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.733902 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 18:20:45.734028 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.733944 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 18:20:45.734028 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.733957 2581 projected.go:194] Error preparing data for projected volume kube-api-access-gbnfq for pod openshift-network-diagnostics/network-check-target-47xwd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:45.734028 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:45.734024 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq podName:e4946d3e-bc1b-4f87-8fb8-db35971cf837 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:46.234005346 +0000 UTC m=+3.038068689 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gbnfq" (UniqueName: "kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq") pod "network-check-target-47xwd" (UID: "e4946d3e-bc1b-4f87-8fb8-db35971cf837") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:45.735851 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.735831 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-88b8b\" (UniqueName: \"kubernetes.io/projected/e6d25e3f-fd86-49df-bcb4-b6ac18c164e1-kube-api-access-88b8b\") pod \"multus-p27q9\" (UID: \"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1\") " pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.735951 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.735874 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw47q\" (UniqueName: \"kubernetes.io/projected/b5ac120e-7d32-45b8-a5d6-57aa386e732f-kube-api-access-mw47q\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:45.736521 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.736494 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnkf8\" (UniqueName: \"kubernetes.io/projected/f0bc6045-fff5-4c2b-a848-e95a54d2aaf8-kube-api-access-gnkf8\") pod \"node-resolver-pqljk\" (UID: \"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8\") " pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.743599 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.742807 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5m6l\" (UniqueName: \"kubernetes.io/projected/c0602c54-63d4-4509-bf26-1bfa53e55b60-kube-api-access-s5m6l\") pod \"multus-additional-cni-plugins-x2fkn\" (UID: \"c0602c54-63d4-4509-bf26-1bfa53e55b60\") " pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.747290 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.747266 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5224k\" (UniqueName: \"kubernetes.io/projected/bbb98e98-69b1-42c6-a9d1-250ac4d831f3-kube-api-access-5224k\") pod \"node-ca-qxft5\" (UID: \"bbb98e98-69b1-42c6-a9d1-250ac4d831f3\") " pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.826698 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826674 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/c32b819b-1ae9-498a-876b-87995a803b2c-konnectivity-ca\") pod \"konnectivity-agent-v2tgg\" (UID: \"c32b819b-1ae9-498a-876b-87995a803b2c\") " pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.826824 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826711 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-lib-modules\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.826824 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826728 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-var-lib-kubelet\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.826824 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826744 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-ovn\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.826824 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826781 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fz55l\" (UniqueName: \"kubernetes.io/projected/a0ad3985-82e5-4fed-89ff-cd6f30087d98-kube-api-access-fz55l\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.826824 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826802 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovnkube-config\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.826824 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826823 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-kubelet\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826843 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-systemd\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826856 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-var-lib-kubelet\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826880 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-lib-modules\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826865 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-host\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826863 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-ovn\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826961 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-kubelet\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826972 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g7h2m\" (UniqueName: \"kubernetes.io/projected/721979eb-ddad-4268-a95a-833cf12d6074-kube-api-access-g7h2m\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.826995 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-systemd\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827014 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-device-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827042 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-host\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827042 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827073 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827076 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-log-socket\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827107 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-cni-bin\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827111 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-device-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827107 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-log-socket\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827135 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-kubernetes\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827161 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827139 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-cni-bin\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827162 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysctl-conf\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827188 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-kubernetes\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827189 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-systemd\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827224 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/721979eb-ddad-4268-a95a-833cf12d6074-tmp\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827227 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/c32b819b-1ae9-498a-876b-87995a803b2c-konnectivity-ca\") pod \"konnectivity-agent-v2tgg\" (UID: \"c32b819b-1ae9-498a-876b-87995a803b2c\") " pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827225 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-run-systemd\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827262 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovn-node-metrics-cert\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827285 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-socket-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827333 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysctl-conf\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827370 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-env-overrides\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827399 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysctl-d\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827425 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-run\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827455 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-etc-selinux\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827478 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-sys-fs\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827502 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827516 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-socket-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.827956 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827528 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-registration-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827557 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0ad3985-82e5-4fed-89ff-cd6f30087d98-host-slash\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827595 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-run-netns\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827618 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-etc-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827651 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827668 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-run-netns\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827704 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-modprobe-d\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827708 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-sys-fs\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827724 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-etc-selinux\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827734 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysctl-d\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827755 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-run\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827770 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-etc-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827785 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-registration-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827793 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0ad3985-82e5-4fed-89ff-cd6f30087d98-host-slash\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827812 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-env-overrides\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827845 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-modprobe-d\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.827906 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysconfig\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.828714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828040 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovnkube-config\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828086 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-etc-sysconfig\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828118 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-slash\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828143 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828169 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-sys\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828196 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-slash\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828211 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-node-log\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828232 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828243 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/c32b819b-1ae9-498a-876b-87995a803b2c-agent-certs\") pod \"konnectivity-agent-v2tgg\" (UID: \"c32b819b-1ae9-498a-876b-87995a803b2c\") " pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828274 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-node-log\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828242 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/721979eb-ddad-4268-a95a-833cf12d6074-sys\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828304 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/a0ad3985-82e5-4fed-89ff-cd6f30087d98-iptables-alerter-script\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828470 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-systemd-units\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828493 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-cni-netd\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828518 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovnkube-script-lib\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828543 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdt2q\" (UniqueName: \"kubernetes.io/projected/938a2b8c-12e5-458f-88a8-d8c54b3148e0-kube-api-access-vdt2q\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828567 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/721979eb-ddad-4268-a95a-833cf12d6074-etc-tuned\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.829651 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828593 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-kubelet-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828617 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-94hqc\" (UniqueName: \"kubernetes.io/projected/90936ce6-5c5f-4f38-9487-0e4631ff77e6-kube-api-access-94hqc\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828645 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-var-lib-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828710 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-var-lib-openvswitch\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828713 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/a0ad3985-82e5-4fed-89ff-cd6f30087d98-iptables-alerter-script\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828773 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-systemd-units\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828808 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/90936ce6-5c5f-4f38-9487-0e4631ff77e6-host-cni-netd\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.828911 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/938a2b8c-12e5-458f-88a8-d8c54b3148e0-kubelet-dir\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.829217 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovnkube-script-lib\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.830472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.830130 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/90936ce6-5c5f-4f38-9487-0e4631ff77e6-ovn-node-metrics-cert\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.830848 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.830511 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/c32b819b-1ae9-498a-876b-87995a803b2c-agent-certs\") pod \"konnectivity-agent-v2tgg\" (UID: \"c32b819b-1ae9-498a-876b-87995a803b2c\") " pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.831011 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.830991 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/721979eb-ddad-4268-a95a-833cf12d6074-tmp\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.831091 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.831044 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/721979eb-ddad-4268-a95a-833cf12d6074-etc-tuned\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.842463 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.842395 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz55l\" (UniqueName: \"kubernetes.io/projected/a0ad3985-82e5-4fed-89ff-cd6f30087d98-kube-api-access-fz55l\") pod \"iptables-alerter-96txh\" (UID: \"a0ad3985-82e5-4fed-89ff-cd6f30087d98\") " pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.842653 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.842619 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7h2m\" (UniqueName: \"kubernetes.io/projected/721979eb-ddad-4268-a95a-833cf12d6074-kube-api-access-g7h2m\") pod \"tuned-zhjmm\" (UID: \"721979eb-ddad-4268-a95a-833cf12d6074\") " pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:45.843226 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.843201 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-94hqc\" (UniqueName: \"kubernetes.io/projected/90936ce6-5c5f-4f38-9487-0e4631ff77e6-kube-api-access-94hqc\") pod \"ovnkube-node-44fqr\" (UID: \"90936ce6-5c5f-4f38-9487-0e4631ff77e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.868787 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.868766 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdt2q\" (UniqueName: \"kubernetes.io/projected/938a2b8c-12e5-458f-88a8-d8c54b3148e0-kube-api-access-vdt2q\") pod \"aws-ebs-csi-driver-node-84n45\" (UID: \"938a2b8c-12e5-458f-88a8-d8c54b3148e0\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.917134 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.917106 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qxft5" Apr 22 18:20:45.925760 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.925741 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" Apr 22 18:20:45.935237 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.935221 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-p27q9" Apr 22 18:20:45.940713 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.940694 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pqljk" Apr 22 18:20:45.947166 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.947145 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-96txh" Apr 22 18:20:45.952743 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.952725 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:20:45.959280 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.959264 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" Apr 22 18:20:45.964856 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.964839 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:20:45.969414 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:45.969399 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" Apr 22 18:20:46.231123 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.231048 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:46.231248 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:46.231187 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:46.231248 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:46.231245 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs podName:b5ac120e-7d32-45b8-a5d6-57aa386e732f nodeName:}" failed. No retries permitted until 2026-04-22 18:20:47.231232545 +0000 UTC m=+4.035295894 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs") pod "network-metrics-daemon-xl5pv" (UID: "b5ac120e-7d32-45b8-a5d6-57aa386e732f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:46.289530 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.289499 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721979eb_ddad_4268_a95a_833cf12d6074.slice/crio-eaf3271e81deaf8948e2afbab6268551cfcd35723e527d1a3baf9a28c3f00f2b WatchSource:0}: Error finding container eaf3271e81deaf8948e2afbab6268551cfcd35723e527d1a3baf9a28c3f00f2b: Status 404 returned error can't find the container with id eaf3271e81deaf8948e2afbab6268551cfcd35723e527d1a3baf9a28c3f00f2b Apr 22 18:20:46.291381 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.291358 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod938a2b8c_12e5_458f_88a8_d8c54b3148e0.slice/crio-5861d7ae174951306c4cc4a9c4a70189cb7fa9c138836186214a3a04bef17926 WatchSource:0}: Error finding container 5861d7ae174951306c4cc4a9c4a70189cb7fa9c138836186214a3a04bef17926: Status 404 returned error can't find the container with id 5861d7ae174951306c4cc4a9c4a70189cb7fa9c138836186214a3a04bef17926 Apr 22 18:20:46.294714 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.294693 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc32b819b_1ae9_498a_876b_87995a803b2c.slice/crio-3bda77b2b8923645dc97671ded3cc87d2ec6a4e3731c2b69beb6d60feef4c184 WatchSource:0}: Error finding container 3bda77b2b8923645dc97671ded3cc87d2ec6a4e3731c2b69beb6d60feef4c184: Status 404 returned error can't find the container with id 3bda77b2b8923645dc97671ded3cc87d2ec6a4e3731c2b69beb6d60feef4c184 Apr 22 18:20:46.295640 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.295620 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6d25e3f_fd86_49df_bcb4_b6ac18c164e1.slice/crio-d154ac55a5228a480a50220b73bef6ae2a228b380637ede5dc3b8af2af45dfeb WatchSource:0}: Error finding container d154ac55a5228a480a50220b73bef6ae2a228b380637ede5dc3b8af2af45dfeb: Status 404 returned error can't find the container with id d154ac55a5228a480a50220b73bef6ae2a228b380637ede5dc3b8af2af45dfeb Apr 22 18:20:46.296685 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.296643 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90936ce6_5c5f_4f38_9487_0e4631ff77e6.slice/crio-7f6358866920c09848da38a3047a8459981c27bfd906b65bde1327ad77ecb133 WatchSource:0}: Error finding container 7f6358866920c09848da38a3047a8459981c27bfd906b65bde1327ad77ecb133: Status 404 returned error can't find the container with id 7f6358866920c09848da38a3047a8459981c27bfd906b65bde1327ad77ecb133 Apr 22 18:20:46.297850 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.297822 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0602c54_63d4_4509_bf26_1bfa53e55b60.slice/crio-1bfcea6658253a49fb96e9503d6d76889124e621a2ae66ec0db11066591f039b WatchSource:0}: Error finding container 1bfcea6658253a49fb96e9503d6d76889124e621a2ae66ec0db11066591f039b: Status 404 returned error can't find the container with id 1bfcea6658253a49fb96e9503d6d76889124e621a2ae66ec0db11066591f039b Apr 22 18:20:46.298716 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.298581 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0bc6045_fff5_4c2b_a848_e95a54d2aaf8.slice/crio-abd34db7d1f2ff365e59f88beac51a7df2cdef4f2679bf943e6fd18e2e44ef26 WatchSource:0}: Error finding container abd34db7d1f2ff365e59f88beac51a7df2cdef4f2679bf943e6fd18e2e44ef26: Status 404 returned error can't find the container with id abd34db7d1f2ff365e59f88beac51a7df2cdef4f2679bf943e6fd18e2e44ef26 Apr 22 18:20:46.299832 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.299808 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0ad3985_82e5_4fed_89ff_cd6f30087d98.slice/crio-1ea77176628fe75d1f6b1559d1eeaea3d4226ad7377ba971566cec4e1f24d7e6 WatchSource:0}: Error finding container 1ea77176628fe75d1f6b1559d1eeaea3d4226ad7377ba971566cec4e1f24d7e6: Status 404 returned error can't find the container with id 1ea77176628fe75d1f6b1559d1eeaea3d4226ad7377ba971566cec4e1f24d7e6 Apr 22 18:20:46.300740 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:20:46.300715 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbb98e98_69b1_42c6_a9d1_250ac4d831f3.slice/crio-3f53b6a33618368e1d77879c586fabf452480aeecce2247866e3f47ab9b71708 WatchSource:0}: Error finding container 3f53b6a33618368e1d77879c586fabf452480aeecce2247866e3f47ab9b71708: Status 404 returned error can't find the container with id 3f53b6a33618368e1d77879c586fabf452480aeecce2247866e3f47ab9b71708 Apr 22 18:20:46.332274 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.332160 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:46.332362 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:46.332303 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 18:20:46.332362 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:46.332320 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 18:20:46.332362 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:46.332329 2581 projected.go:194] Error preparing data for projected volume kube-api-access-gbnfq for pod openshift-network-diagnostics/network-check-target-47xwd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:46.332475 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:46.332372 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq podName:e4946d3e-bc1b-4f87-8fb8-db35971cf837 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:47.33235769 +0000 UTC m=+4.136421033 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-gbnfq" (UniqueName: "kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq") pod "network-check-target-47xwd" (UID: "e4946d3e-bc1b-4f87-8fb8-db35971cf837") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:46.651430 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.651348 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 18:15:44 +0000 UTC" deadline="2028-02-06 18:56:29.528226702 +0000 UTC" Apr 22 18:20:46.651430 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.651375 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="15720h35m42.876853751s" Apr 22 18:20:46.757946 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.757500 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" event={"ID":"b128912d2b3d8968717857155603d30b","Type":"ContainerStarted","Data":"8ac9592f8c376d64ab9e5b5db8a237d96f1e6a6c4a951ca46a4260c66349869e"} Apr 22 18:20:46.767287 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.766304 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-96txh" event={"ID":"a0ad3985-82e5-4fed-89ff-cd6f30087d98","Type":"ContainerStarted","Data":"1ea77176628fe75d1f6b1559d1eeaea3d4226ad7377ba971566cec4e1f24d7e6"} Apr 22 18:20:46.770954 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.770765 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pqljk" event={"ID":"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8","Type":"ContainerStarted","Data":"abd34db7d1f2ff365e59f88beac51a7df2cdef4f2679bf943e6fd18e2e44ef26"} Apr 22 18:20:46.782669 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.782151 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-141-172.ec2.internal" podStartSLOduration=1.7821352940000001 podStartE2EDuration="1.782135294s" podCreationTimestamp="2026-04-22 18:20:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 18:20:46.781542375 +0000 UTC m=+3.585605738" watchObservedRunningTime="2026-04-22 18:20:46.782135294 +0000 UTC m=+3.586198661" Apr 22 18:20:46.782669 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.782221 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerStarted","Data":"1bfcea6658253a49fb96e9503d6d76889124e621a2ae66ec0db11066591f039b"} Apr 22 18:20:46.784311 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.784270 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p27q9" event={"ID":"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1","Type":"ContainerStarted","Data":"d154ac55a5228a480a50220b73bef6ae2a228b380637ede5dc3b8af2af45dfeb"} Apr 22 18:20:46.788597 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.788373 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" event={"ID":"938a2b8c-12e5-458f-88a8-d8c54b3148e0","Type":"ContainerStarted","Data":"5861d7ae174951306c4cc4a9c4a70189cb7fa9c138836186214a3a04bef17926"} Apr 22 18:20:46.796321 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.796254 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qxft5" event={"ID":"bbb98e98-69b1-42c6-a9d1-250ac4d831f3","Type":"ContainerStarted","Data":"3f53b6a33618368e1d77879c586fabf452480aeecce2247866e3f47ab9b71708"} Apr 22 18:20:46.798410 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.798389 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"7f6358866920c09848da38a3047a8459981c27bfd906b65bde1327ad77ecb133"} Apr 22 18:20:46.804808 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.804786 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-v2tgg" event={"ID":"c32b819b-1ae9-498a-876b-87995a803b2c","Type":"ContainerStarted","Data":"3bda77b2b8923645dc97671ded3cc87d2ec6a4e3731c2b69beb6d60feef4c184"} Apr 22 18:20:46.806822 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:46.806799 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" event={"ID":"721979eb-ddad-4268-a95a-833cf12d6074","Type":"ContainerStarted","Data":"eaf3271e81deaf8948e2afbab6268551cfcd35723e527d1a3baf9a28c3f00f2b"} Apr 22 18:20:47.240529 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:47.240443 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:47.240693 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.240631 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:47.240693 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.240692 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs podName:b5ac120e-7d32-45b8-a5d6-57aa386e732f nodeName:}" failed. No retries permitted until 2026-04-22 18:20:49.240674355 +0000 UTC m=+6.044737700 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs") pod "network-metrics-daemon-xl5pv" (UID: "b5ac120e-7d32-45b8-a5d6-57aa386e732f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:47.341460 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:47.340832 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:47.341460 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.341005 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 18:20:47.341460 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.341024 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 18:20:47.341460 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.341037 2581 projected.go:194] Error preparing data for projected volume kube-api-access-gbnfq for pod openshift-network-diagnostics/network-check-target-47xwd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:47.341460 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.341091 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq podName:e4946d3e-bc1b-4f87-8fb8-db35971cf837 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:49.341073493 +0000 UTC m=+6.145136841 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-gbnfq" (UniqueName: "kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq") pod "network-check-target-47xwd" (UID: "e4946d3e-bc1b-4f87-8fb8-db35971cf837") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:47.741644 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:47.741102 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:47.741644 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.741215 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:47.741644 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:47.741102 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:47.741644 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:47.741326 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:20:47.820021 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:47.819989 2581 generic.go:358] "Generic (PLEG): container finished" podID="060906b545f652629bb1b4f63cd4a0c8" containerID="b7b8e8c63d2412b6843318fe13c05365aa2b2c98f5765fa2ef2034aadad822f0" exitCode=0 Apr 22 18:20:47.820888 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:47.820857 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" event={"ID":"060906b545f652629bb1b4f63cd4a0c8","Type":"ContainerDied","Data":"b7b8e8c63d2412b6843318fe13c05365aa2b2c98f5765fa2ef2034aadad822f0"} Apr 22 18:20:48.827004 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:48.826968 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" event={"ID":"060906b545f652629bb1b4f63cd4a0c8","Type":"ContainerStarted","Data":"9de72405dcce9f67eb9a55f719b414c361984ec170bd87f7034d4010c82c78d9"} Apr 22 18:20:49.256181 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:49.256110 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:49.256324 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.256283 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:49.256386 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.256349 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs podName:b5ac120e-7d32-45b8-a5d6-57aa386e732f nodeName:}" failed. No retries permitted until 2026-04-22 18:20:53.256330579 +0000 UTC m=+10.060393945 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs") pod "network-metrics-daemon-xl5pv" (UID: "b5ac120e-7d32-45b8-a5d6-57aa386e732f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:49.357515 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:49.357476 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:49.357701 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.357628 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 18:20:49.357701 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.357647 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 18:20:49.357701 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.357657 2581 projected.go:194] Error preparing data for projected volume kube-api-access-gbnfq for pod openshift-network-diagnostics/network-check-target-47xwd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:49.357701 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.357700 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq podName:e4946d3e-bc1b-4f87-8fb8-db35971cf837 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:53.357687041 +0000 UTC m=+10.161750398 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-gbnfq" (UniqueName: "kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq") pod "network-check-target-47xwd" (UID: "e4946d3e-bc1b-4f87-8fb8-db35971cf837") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:49.740142 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:49.740045 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:49.740300 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.740167 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:20:49.741863 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:49.741834 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:49.741998 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:49.741946 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:50.564232 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.564179 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-141-172.ec2.internal" podStartSLOduration=5.5641595200000005 podStartE2EDuration="5.56415952s" podCreationTimestamp="2026-04-22 18:20:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 18:20:48.840548608 +0000 UTC m=+5.644611975" watchObservedRunningTime="2026-04-22 18:20:50.56415952 +0000 UTC m=+7.368222887" Apr 22 18:20:50.564574 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.564411 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-x5xw7"] Apr 22 18:20:50.572543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.572520 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.572777 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:50.572755 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:20:50.670800 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.670765 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f0d1648a-0bbf-4049-b68e-82046b4578f0-dbus\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.670976 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.670815 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.670976 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.670866 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f0d1648a-0bbf-4049-b68e-82046b4578f0-kubelet-config\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.772086 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.772055 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f0d1648a-0bbf-4049-b68e-82046b4578f0-kubelet-config\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.772216 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.772134 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f0d1648a-0bbf-4049-b68e-82046b4578f0-dbus\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.772216 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.772163 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.772300 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:50.772268 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:50.772340 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:50.772328 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret podName:f0d1648a-0bbf-4049-b68e-82046b4578f0 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:51.272310261 +0000 UTC m=+8.076373615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret") pod "global-pull-secret-syncer-x5xw7" (UID: "f0d1648a-0bbf-4049-b68e-82046b4578f0") : object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:50.772573 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.772561 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f0d1648a-0bbf-4049-b68e-82046b4578f0-kubelet-config\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:50.772702 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:50.772689 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f0d1648a-0bbf-4049-b68e-82046b4578f0-dbus\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:51.277247 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:51.277213 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:51.277408 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:51.277335 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:51.277408 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:51.277379 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret podName:f0d1648a-0bbf-4049-b68e-82046b4578f0 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:52.277366358 +0000 UTC m=+9.081429701 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret") pod "global-pull-secret-syncer-x5xw7" (UID: "f0d1648a-0bbf-4049-b68e-82046b4578f0") : object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:51.743542 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:51.743449 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:51.743542 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:51.743453 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:51.744018 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:51.743552 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:51.744018 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:51.743656 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:20:52.286836 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:52.286796 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:52.287043 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:52.287005 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:52.287103 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:52.287073 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret podName:f0d1648a-0bbf-4049-b68e-82046b4578f0 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:54.28705442 +0000 UTC m=+11.091117768 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret") pod "global-pull-secret-syncer-x5xw7" (UID: "f0d1648a-0bbf-4049-b68e-82046b4578f0") : object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:52.739368 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:52.739294 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:52.739527 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:52.739403 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:20:53.294065 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:53.294032 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:53.294510 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.294160 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:53.294510 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.294217 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs podName:b5ac120e-7d32-45b8-a5d6-57aa386e732f nodeName:}" failed. No retries permitted until 2026-04-22 18:21:01.294198089 +0000 UTC m=+18.098261439 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs") pod "network-metrics-daemon-xl5pv" (UID: "b5ac120e-7d32-45b8-a5d6-57aa386e732f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:20:53.394870 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:53.394839 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:53.395073 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.395029 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 18:20:53.395073 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.395046 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 18:20:53.395073 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.395058 2581 projected.go:194] Error preparing data for projected volume kube-api-access-gbnfq for pod openshift-network-diagnostics/network-check-target-47xwd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:53.395212 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.395113 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq podName:e4946d3e-bc1b-4f87-8fb8-db35971cf837 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:01.395094106 +0000 UTC m=+18.199157458 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-gbnfq" (UniqueName: "kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq") pod "network-check-target-47xwd" (UID: "e4946d3e-bc1b-4f87-8fb8-db35971cf837") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:20:53.742737 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:53.742666 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:53.742875 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.742778 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:20:53.743137 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:53.743120 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:53.743239 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:53.743197 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:54.301711 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:54.301677 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:54.302153 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:54.301866 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:54.302153 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:54.301934 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret podName:f0d1648a-0bbf-4049-b68e-82046b4578f0 nodeName:}" failed. No retries permitted until 2026-04-22 18:20:58.301901907 +0000 UTC m=+15.105965256 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret") pod "global-pull-secret-syncer-x5xw7" (UID: "f0d1648a-0bbf-4049-b68e-82046b4578f0") : object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:54.739417 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:54.739335 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:54.739576 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:54.739451 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:20:55.739136 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:55.739064 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:55.739530 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:55.739061 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:55.739530 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:55.739177 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:20:55.739530 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:55.739256 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:56.739539 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:56.739509 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:56.739911 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:56.739602 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:20:57.739296 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:57.739269 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:57.739485 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:57.739269 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:57.739485 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:57.739396 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:20:57.739485 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:57.739475 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:58.331629 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:58.331588 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:58.331798 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:58.331704 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:58.331798 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:58.331754 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret podName:f0d1648a-0bbf-4049-b68e-82046b4578f0 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:06.331742095 +0000 UTC m=+23.135805438 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret") pod "global-pull-secret-syncer-x5xw7" (UID: "f0d1648a-0bbf-4049-b68e-82046b4578f0") : object "kube-system"/"original-pull-secret" not registered Apr 22 18:20:58.739189 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:58.739106 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:20:58.739349 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:58.739242 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:20:59.739154 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:59.739120 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:20:59.739154 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:20:59.739154 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:20:59.739641 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:59.739254 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:20:59.739641 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:20:59.739424 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:00.739653 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:00.739625 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:00.740018 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:00.739719 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:01.359164 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:01.359132 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:01.359362 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.359291 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:21:01.359423 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.359368 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs podName:b5ac120e-7d32-45b8-a5d6-57aa386e732f nodeName:}" failed. No retries permitted until 2026-04-22 18:21:17.359349909 +0000 UTC m=+34.163413268 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs") pod "network-metrics-daemon-xl5pv" (UID: "b5ac120e-7d32-45b8-a5d6-57aa386e732f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:21:01.459985 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:01.459948 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:01.460161 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.460051 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 18:21:01.460161 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.460071 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 18:21:01.460161 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.460080 2581 projected.go:194] Error preparing data for projected volume kube-api-access-gbnfq for pod openshift-network-diagnostics/network-check-target-47xwd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:21:01.460161 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.460138 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq podName:e4946d3e-bc1b-4f87-8fb8-db35971cf837 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:17.460120602 +0000 UTC m=+34.264183960 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-gbnfq" (UniqueName: "kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq") pod "network-check-target-47xwd" (UID: "e4946d3e-bc1b-4f87-8fb8-db35971cf837") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:21:01.739150 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:01.739066 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:01.739294 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:01.739070 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:01.739294 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.739209 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:01.739384 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:01.739310 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:02.739515 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:02.739487 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:02.739865 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:02.739574 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:03.739911 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.739630 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:03.740577 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.739715 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:03.740577 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:03.740013 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:03.740577 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:03.740079 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:03.852572 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.852546 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qxft5" event={"ID":"bbb98e98-69b1-42c6-a9d1-250ac4d831f3","Type":"ContainerStarted","Data":"28166b4a5e6580577fe7d4f008f1e86da2b18569e763f6d09d7edb83125bf3f4"} Apr 22 18:21:03.854358 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.854341 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-acl-logging/0.log" Apr 22 18:21:03.854634 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.854614 2581 generic.go:358] "Generic (PLEG): container finished" podID="90936ce6-5c5f-4f38-9487-0e4631ff77e6" containerID="195fa2e24b1593c3e2f8d78c97a5cd7a865261828fd2d6ca50e49656ae5cc062" exitCode=1 Apr 22 18:21:03.854715 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.854690 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"5129d50cbe18f012a8d2ff2fb227ac772fb41544522624b67b10a7d7d8a252f4"} Apr 22 18:21:03.854771 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.854719 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"56c7d2a589962a0ca1c267b94b8dc66692cfb3a53e9b65b112ee839ded13cb17"} Apr 22 18:21:03.854771 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.854734 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"005451c0024791358a61d5478e498fccaebc50f67e119c4b296fa2d7dae2a0cc"} Apr 22 18:21:03.854771 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.854748 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerDied","Data":"195fa2e24b1593c3e2f8d78c97a5cd7a865261828fd2d6ca50e49656ae5cc062"} Apr 22 18:21:03.854771 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.854761 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"226205d2d62d7151eb55ee741c901e44e63d035242d84290f9a2a971667c18cd"} Apr 22 18:21:03.858688 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.858669 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-v2tgg" event={"ID":"c32b819b-1ae9-498a-876b-87995a803b2c","Type":"ContainerStarted","Data":"db7c4116a952b3d34af7236d066a95c5c19abdf97d1dba819d6bdc51a610c4ba"} Apr 22 18:21:03.859860 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.859841 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" event={"ID":"721979eb-ddad-4268-a95a-833cf12d6074","Type":"ContainerStarted","Data":"2623019a65a2aba19d4bc0be4ae06b189123987c41ac10f9a85f57682467c459"} Apr 22 18:21:03.860889 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.860864 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pqljk" event={"ID":"f0bc6045-fff5-4c2b-a848-e95a54d2aaf8","Type":"ContainerStarted","Data":"61380161c2f205ba03c418ab50e6362ca2d4f9a6fccf2553f11c01c20b4ffd39"} Apr 22 18:21:03.862109 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.862088 2581 generic.go:358] "Generic (PLEG): container finished" podID="c0602c54-63d4-4509-bf26-1bfa53e55b60" containerID="93aede3947b9e2b9872ea29645389fda1c26133a08284ca523fdd1038e463ce2" exitCode=0 Apr 22 18:21:03.862194 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.862139 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerDied","Data":"93aede3947b9e2b9872ea29645389fda1c26133a08284ca523fdd1038e463ce2"} Apr 22 18:21:03.863475 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.863332 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p27q9" event={"ID":"e6d25e3f-fd86-49df-bcb4-b6ac18c164e1","Type":"ContainerStarted","Data":"e025735be1a2f905dc370e2743861dcf6c3ecaf70a80c1f4384b08a9d084dda2"} Apr 22 18:21:03.864501 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.864480 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" event={"ID":"938a2b8c-12e5-458f-88a8-d8c54b3148e0","Type":"ContainerStarted","Data":"1734891258d421876f14898202c2f62fe03592cbd86c5c523beadcc353036216"} Apr 22 18:21:03.891152 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.891107 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-p27q9" podStartSLOduration=4.015005322 podStartE2EDuration="20.89109675s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.297693273 +0000 UTC m=+3.101756629" lastFinishedPulling="2026-04-22 18:21:03.173784697 +0000 UTC m=+19.977848057" observedRunningTime="2026-04-22 18:21:03.89108782 +0000 UTC m=+20.695151181" watchObservedRunningTime="2026-04-22 18:21:03.89109675 +0000 UTC m=+20.695160111" Apr 22 18:21:03.891702 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.891662 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-qxft5" podStartSLOduration=4.181798848 podStartE2EDuration="20.891652715s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.302526209 +0000 UTC m=+3.106589566" lastFinishedPulling="2026-04-22 18:21:03.012380086 +0000 UTC m=+19.816443433" observedRunningTime="2026-04-22 18:21:03.873202324 +0000 UTC m=+20.677265688" watchObservedRunningTime="2026-04-22 18:21:03.891652715 +0000 UTC m=+20.695716085" Apr 22 18:21:03.976346 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:03.976300 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-pqljk" podStartSLOduration=4.104066083 podStartE2EDuration="20.976282943s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.301777643 +0000 UTC m=+3.105840998" lastFinishedPulling="2026-04-22 18:21:03.173994499 +0000 UTC m=+19.978057858" observedRunningTime="2026-04-22 18:21:03.948541904 +0000 UTC m=+20.752605274" watchObservedRunningTime="2026-04-22 18:21:03.976282943 +0000 UTC m=+20.780346308" Apr 22 18:21:04.016986 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:04.016945 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-v2tgg" podStartSLOduration=4.325210659 podStartE2EDuration="21.016934719s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.296629033 +0000 UTC m=+3.100692377" lastFinishedPulling="2026-04-22 18:21:02.988353079 +0000 UTC m=+19.792416437" observedRunningTime="2026-04-22 18:21:03.976008752 +0000 UTC m=+20.780072127" watchObservedRunningTime="2026-04-22 18:21:04.016934719 +0000 UTC m=+20.820998078" Apr 22 18:21:04.017157 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:04.017134 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-zhjmm" podStartSLOduration=4.166106918 podStartE2EDuration="21.01712659s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.291271503 +0000 UTC m=+3.095334848" lastFinishedPulling="2026-04-22 18:21:03.142291176 +0000 UTC m=+19.946354520" observedRunningTime="2026-04-22 18:21:04.017112584 +0000 UTC m=+20.821175971" watchObservedRunningTime="2026-04-22 18:21:04.01712659 +0000 UTC m=+20.821189955" Apr 22 18:21:04.739737 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:04.739698 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:04.739857 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:04.739798 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:04.868610 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:04.868584 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-acl-logging/0.log" Apr 22 18:21:04.868972 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:04.868955 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"eb89682e79f6e8ac0680deff8ee2323209ec8ab2954729d494260e1cea11c0ce"} Apr 22 18:21:04.870249 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:04.870225 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-96txh" event={"ID":"a0ad3985-82e5-4fed-89ff-cd6f30087d98","Type":"ContainerStarted","Data":"4b7a6f24065800b953e5218675d97bdb9f12d89abb8a6bf702b15d81e8056107"} Apr 22 18:21:04.902569 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:04.902466 2581 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 22 18:21:05.691109 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:05.690968 2581 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-22T18:21:04.902483394Z","UUID":"56cc07ae-30f3-4e4c-917a-34a0cb275383","Handler":null,"Name":"","Endpoint":""} Apr 22 18:21:05.696288 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:05.696258 2581 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 22 18:21:05.696288 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:05.696287 2581 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 22 18:21:05.739812 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:05.739783 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:05.739961 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:05.739899 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:05.740180 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:05.740162 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:05.740256 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:05.740241 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:05.874119 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:05.874091 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" event={"ID":"938a2b8c-12e5-458f-88a8-d8c54b3148e0","Type":"ContainerStarted","Data":"e2ed16e401a559622857df7c7356fbce5761f2c142c958c8bce539de3742d197"} Apr 22 18:21:06.185778 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.185754 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:21:06.186340 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.186321 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:21:06.203815 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.203772 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-96txh" podStartSLOduration=6.49361789 podStartE2EDuration="23.203757047s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.302179321 +0000 UTC m=+3.106242668" lastFinishedPulling="2026-04-22 18:21:03.012318469 +0000 UTC m=+19.816381825" observedRunningTime="2026-04-22 18:21:04.889219505 +0000 UTC m=+21.693282870" watchObservedRunningTime="2026-04-22 18:21:06.203757047 +0000 UTC m=+23.007820412" Apr 22 18:21:06.399257 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.399068 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:06.399426 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:06.399206 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 18:21:06.399426 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:06.399331 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret podName:f0d1648a-0bbf-4049-b68e-82046b4578f0 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:22.399311958 +0000 UTC m=+39.203375302 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret") pod "global-pull-secret-syncer-x5xw7" (UID: "f0d1648a-0bbf-4049-b68e-82046b4578f0") : object "kube-system"/"original-pull-secret" not registered Apr 22 18:21:06.739449 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.739420 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:06.739600 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:06.739526 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:06.878369 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.878332 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" event={"ID":"938a2b8c-12e5-458f-88a8-d8c54b3148e0","Type":"ContainerStarted","Data":"332830fb2188eb2ba5f2c4aab8430e8989e85a6aa8b0ebf7fb06f7bd919ae0b2"} Apr 22 18:21:06.881160 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.881136 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-acl-logging/0.log" Apr 22 18:21:06.881514 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:06.881488 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"201c3e96e8e916d529f635a250154d66485a71865748ec34084987e6bfe138b7"} Apr 22 18:21:07.742851 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:07.742830 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:07.743062 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:07.742965 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:07.743062 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:07.742972 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:07.743156 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:07.743067 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:07.743156 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:07.743109 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:07.743249 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:07.743179 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:08.886831 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.886614 2581 generic.go:358] "Generic (PLEG): container finished" podID="c0602c54-63d4-4509-bf26-1bfa53e55b60" containerID="155895291f5b6b9102e2c8c60f67b67766937b25b7186fc598e14962c0726e1b" exitCode=0 Apr 22 18:21:08.886831 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.886690 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerDied","Data":"155895291f5b6b9102e2c8c60f67b67766937b25b7186fc598e14962c0726e1b"} Apr 22 18:21:08.890212 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.890191 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-acl-logging/0.log" Apr 22 18:21:08.890550 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.890530 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"2ae7a0e2536c435d68b7ea24b9ef5ee7853d2fda991991c7d5806c1eeabd7dd2"} Apr 22 18:21:08.890814 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.890794 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:21:08.890853 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.890828 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:21:08.891061 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.891045 2581 scope.go:117] "RemoveContainer" containerID="195fa2e24b1593c3e2f8d78c97a5cd7a865261828fd2d6ca50e49656ae5cc062" Apr 22 18:21:08.911633 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.911599 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-84n45" podStartSLOduration=6.121745833 podStartE2EDuration="25.911589423s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.29343334 +0000 UTC m=+3.097496686" lastFinishedPulling="2026-04-22 18:21:06.08327692 +0000 UTC m=+22.887340276" observedRunningTime="2026-04-22 18:21:06.89725962 +0000 UTC m=+23.701322984" watchObservedRunningTime="2026-04-22 18:21:08.911589423 +0000 UTC m=+25.715652830" Apr 22 18:21:08.913169 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:08.913152 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:21:09.450523 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.450497 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:21:09.450688 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.450599 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 22 18:21:09.450980 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.450966 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-v2tgg" Apr 22 18:21:09.739229 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.739159 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:09.739344 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:09.739250 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:09.739344 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.739272 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:09.739344 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.739295 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:09.739449 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:09.739346 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:09.739449 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:09.739432 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:09.895454 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.895428 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-acl-logging/0.log" Apr 22 18:21:09.895812 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.895793 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" event={"ID":"90936ce6-5c5f-4f38-9487-0e4631ff77e6","Type":"ContainerStarted","Data":"b99f2bc854cd3469ccbc4427a53b346913d275a39b6d05397c25fec335ec6643"} Apr 22 18:21:09.896156 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.896133 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:21:09.909693 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.909672 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:21:09.932178 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:09.932132 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" podStartSLOduration=10.011340181 podStartE2EDuration="26.9321157s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.299062408 +0000 UTC m=+3.103125752" lastFinishedPulling="2026-04-22 18:21:03.219837915 +0000 UTC m=+20.023901271" observedRunningTime="2026-04-22 18:21:09.931727199 +0000 UTC m=+26.735790564" watchObservedRunningTime="2026-04-22 18:21:09.9321157 +0000 UTC m=+26.736179066" Apr 22 18:21:10.305660 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.305587 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-x5xw7"] Apr 22 18:21:10.305802 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.305711 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:10.305802 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:10.305787 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:10.307543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.307517 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-47xwd"] Apr 22 18:21:10.307653 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.307619 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:10.307704 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:10.307687 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:10.314708 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.314685 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xl5pv"] Apr 22 18:21:10.314812 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.314798 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:10.314968 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:10.314944 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:10.899627 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.899598 2581 generic.go:358] "Generic (PLEG): container finished" podID="c0602c54-63d4-4509-bf26-1bfa53e55b60" containerID="288939d4325ed4854254b61db400d46ff3355f79c5048611c1f87de49e2e5fbe" exitCode=0 Apr 22 18:21:10.899983 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:10.899689 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerDied","Data":"288939d4325ed4854254b61db400d46ff3355f79c5048611c1f87de49e2e5fbe"} Apr 22 18:21:11.739466 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:11.739300 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:11.739629 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:11.739300 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:11.739629 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:11.739534 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:11.739629 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:11.739306 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:11.739733 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:11.739625 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:11.739733 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:11.739714 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:12.906447 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:12.906419 2581 generic.go:358] "Generic (PLEG): container finished" podID="c0602c54-63d4-4509-bf26-1bfa53e55b60" containerID="5b605c6162fb74bc9559813706174e867ff155b549c6ab32e8d0e2aaa2a4c4d3" exitCode=0 Apr 22 18:21:12.906800 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:12.906478 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerDied","Data":"5b605c6162fb74bc9559813706174e867ff155b549c6ab32e8d0e2aaa2a4c4d3"} Apr 22 18:21:13.740158 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:13.740129 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:13.740308 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:13.740205 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:13.740502 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:13.740485 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:13.740586 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:13.740567 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:13.740646 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:13.740609 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:13.740694 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:13.740657 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:14.834613 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:14.834548 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-pqljk_f0bc6045-fff5-4c2b-a848-e95a54d2aaf8/dns-node-resolver/0.log" Apr 22 18:21:15.416173 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:15.416139 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-qxft5_bbb98e98-69b1-42c6-a9d1-250ac4d831f3/node-ca/0.log" Apr 22 18:21:15.739196 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:15.739116 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:15.739196 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:15.739157 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:15.739390 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:15.739237 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:15.739390 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:15.739319 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:15.739390 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:15.739229 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:15.739624 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:15.739397 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:17.387542 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:17.387509 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:17.388058 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.387641 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:21:17.388058 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.387712 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs podName:b5ac120e-7d32-45b8-a5d6-57aa386e732f nodeName:}" failed. No retries permitted until 2026-04-22 18:21:49.387693201 +0000 UTC m=+66.191756557 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs") pod "network-metrics-daemon-xl5pv" (UID: "b5ac120e-7d32-45b8-a5d6-57aa386e732f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 18:21:17.488834 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:17.488802 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:17.489025 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.489004 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 18:21:17.489082 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.489029 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 18:21:17.489082 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.489043 2581 projected.go:194] Error preparing data for projected volume kube-api-access-gbnfq for pod openshift-network-diagnostics/network-check-target-47xwd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:21:17.489155 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.489094 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq podName:e4946d3e-bc1b-4f87-8fb8-db35971cf837 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:49.489079558 +0000 UTC m=+66.293142902 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-gbnfq" (UniqueName: "kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq") pod "network-check-target-47xwd" (UID: "e4946d3e-bc1b-4f87-8fb8-db35971cf837") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 18:21:17.739833 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:17.739732 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:17.739833 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:17.739733 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:17.740061 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.739862 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:17.740061 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:17.739740 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:17.740061 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.739961 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:17.740061 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:17.740035 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:18.919876 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:18.919843 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerStarted","Data":"7daf2cd5efbb1c43592e56ead546d09e8ff40471b3b01f3f9c6ff981e8b332c8"} Apr 22 18:21:19.739786 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:19.739759 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:19.739973 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:19.739759 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:19.739973 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:19.739853 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:19.739973 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:19.739762 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:19.739973 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:19.739937 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:19.740144 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:19.740021 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:19.923840 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:19.923812 2581 generic.go:358] "Generic (PLEG): container finished" podID="c0602c54-63d4-4509-bf26-1bfa53e55b60" containerID="7daf2cd5efbb1c43592e56ead546d09e8ff40471b3b01f3f9c6ff981e8b332c8" exitCode=0 Apr 22 18:21:19.924250 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:19.923847 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerDied","Data":"7daf2cd5efbb1c43592e56ead546d09e8ff40471b3b01f3f9c6ff981e8b332c8"} Apr 22 18:21:20.928191 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:20.928160 2581 generic.go:358] "Generic (PLEG): container finished" podID="c0602c54-63d4-4509-bf26-1bfa53e55b60" containerID="df8857bc457e27df0f530b1ebc09901fdfc9e8587cc32e77126807042e0d593b" exitCode=0 Apr 22 18:21:20.928539 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:20.928202 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerDied","Data":"df8857bc457e27df0f530b1ebc09901fdfc9e8587cc32e77126807042e0d593b"} Apr 22 18:21:21.739509 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:21.739345 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:21.739681 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:21.739576 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:21.739681 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:21.739359 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:21.739681 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:21.739643 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:21.739681 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:21.739346 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:21.739828 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:21.739711 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:21.932046 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:21.932016 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" event={"ID":"c0602c54-63d4-4509-bf26-1bfa53e55b60","Type":"ContainerStarted","Data":"54383596f6f355ee1d40d91c2e8d90008325a94a0674b1eb3c7a1eacec0a78ae"} Apr 22 18:21:21.956793 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:21.956755 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-x2fkn" podStartSLOduration=6.524556884 podStartE2EDuration="38.956745216s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:20:46.301456267 +0000 UTC m=+3.105519618" lastFinishedPulling="2026-04-22 18:21:18.733644604 +0000 UTC m=+35.537707950" observedRunningTime="2026-04-22 18:21:21.954980583 +0000 UTC m=+38.759043949" watchObservedRunningTime="2026-04-22 18:21:21.956745216 +0000 UTC m=+38.760808581" Apr 22 18:21:22.427289 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:22.427258 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:22.427458 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:22.427362 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 18:21:22.427458 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:22.427420 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret podName:f0d1648a-0bbf-4049-b68e-82046b4578f0 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:54.427403225 +0000 UTC m=+71.231466585 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret") pod "global-pull-secret-syncer-x5xw7" (UID: "f0d1648a-0bbf-4049-b68e-82046b4578f0") : object "kube-system"/"original-pull-secret" not registered Apr 22 18:21:23.738909 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:23.738882 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:23.739687 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:23.739674 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:23.739761 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:23.739741 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:23.739761 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:23.739755 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:23.739867 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:23.739827 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:23.739941 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:23.739897 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:25.739440 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:25.739407 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:25.739440 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:25.739433 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:25.740034 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:25.739407 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:25.740034 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:25.739507 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:25.740034 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:25.739566 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:25.740034 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:25.739617 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:27.739118 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:27.739089 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:27.739560 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:27.739188 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:27.739560 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:27.739202 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:27.739560 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:27.739272 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:27.739560 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:27.739308 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:27.739560 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:27.739374 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:29.739155 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:29.739124 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:29.739155 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:29.739152 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:29.739568 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:29.739152 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:29.739568 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:29.739210 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:29.739568 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:29.739268 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:29.739568 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:29.739349 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:31.739854 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:31.739822 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:31.739854 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:31.739836 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:31.740357 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:31.739819 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:31.740357 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:31.739913 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:31.740357 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:31.740006 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:31.740357 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:31.740077 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:33.745412 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:33.745384 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:33.745932 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:33.745384 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:33.745932 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:33.745494 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-47xwd" podUID="e4946d3e-bc1b-4f87-8fb8-db35971cf837" Apr 22 18:21:33.745932 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:33.745527 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:33.745932 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:33.745617 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-x5xw7" podUID="f0d1648a-0bbf-4049-b68e-82046b4578f0" Apr 22 18:21:33.745932 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:33.745809 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xl5pv" podUID="b5ac120e-7d32-45b8-a5d6-57aa386e732f" Apr 22 18:21:34.016440 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.016410 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-141-172.ec2.internal" event="NodeReady" Apr 22 18:21:34.016552 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.016514 2581 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 22 18:21:34.059058 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.059034 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-f6646996-g48x4"] Apr 22 18:21:34.093813 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.093787 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-f6646996-g48x4"] Apr 22 18:21:34.093913 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.093818 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.096815 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.096799 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-4bvf4\"" Apr 22 18:21:34.096901 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.096853 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 22 18:21:34.096961 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.096942 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 22 18:21:34.098973 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.098953 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 22 18:21:34.102494 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.102478 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 22 18:21:34.116848 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.116828 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/0326059a-7ea7-44ae-bc43-7619240760e7-image-registry-private-configuration\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.116945 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.116868 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-bound-sa-token\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.116945 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.116890 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhxzz\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-kube-api-access-jhxzz\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.117014 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.116958 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-registry-tls\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.117014 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.116973 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0326059a-7ea7-44ae-bc43-7619240760e7-registry-certificates\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.117014 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.117000 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0326059a-7ea7-44ae-bc43-7619240760e7-trusted-ca\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.117108 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.117015 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0326059a-7ea7-44ae-bc43-7619240760e7-installation-pull-secrets\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.117108 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.117033 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0326059a-7ea7-44ae-bc43-7619240760e7-ca-trust-extracted\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.124023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.124006 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-57ft4"] Apr 22 18:21:34.137687 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.137665 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b9bkd"] Apr 22 18:21:34.137836 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.137815 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.140867 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.140851 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-69x95\"" Apr 22 18:21:34.141201 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.141187 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 22 18:21:34.141281 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.141245 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 22 18:21:34.159488 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.159470 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-57ft4"] Apr 22 18:21:34.159488 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.159490 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b9bkd"] Apr 22 18:21:34.159602 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.159499 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-sh9jw"] Apr 22 18:21:34.159602 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.159591 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.162879 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.162851 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 22 18:21:34.163810 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.163788 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 22 18:21:34.164026 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.163940 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-dwl6t\"" Apr 22 18:21:34.164026 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.164018 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 22 18:21:34.177744 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.177727 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-sh9jw"] Apr 22 18:21:34.177825 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.177812 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.180746 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.180685 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 22 18:21:34.181006 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.180990 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 22 18:21:34.181507 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.181485 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-592td\"" Apr 22 18:21:34.181592 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.181485 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 22 18:21:34.181592 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.181571 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 22 18:21:34.217497 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217479 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47l6r\" (UniqueName: \"kubernetes.io/projected/629aaa70-005e-4af9-9ce5-ca15d667f5e7-kube-api-access-47l6r\") pod \"ingress-canary-b9bkd\" (UID: \"629aaa70-005e-4af9-9ce5-ca15d667f5e7\") " pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.217597 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217514 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-registry-tls\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.217597 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217536 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0326059a-7ea7-44ae-bc43-7619240760e7-registry-certificates\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.217597 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217553 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/629aaa70-005e-4af9-9ce5-ca15d667f5e7-cert\") pod \"ingress-canary-b9bkd\" (UID: \"629aaa70-005e-4af9-9ce5-ca15d667f5e7\") " pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.217743 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217595 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-config-volume\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.217743 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217655 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0326059a-7ea7-44ae-bc43-7619240760e7-trusted-ca\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.217743 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217685 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0326059a-7ea7-44ae-bc43-7619240760e7-installation-pull-secrets\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.217885 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217745 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0326059a-7ea7-44ae-bc43-7619240760e7-ca-trust-extracted\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.217885 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217777 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/8315b63f-0468-47cb-b528-7534d99126a4-crio-socket\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.217885 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217845 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/8315b63f-0468-47cb-b528-7534d99126a4-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.218062 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217897 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/0326059a-7ea7-44ae-bc43-7619240760e7-image-registry-private-configuration\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.218062 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.217974 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-metrics-tls\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.218062 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218024 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q7fq\" (UniqueName: \"kubernetes.io/projected/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-kube-api-access-8q7fq\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.218211 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218064 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/8315b63f-0468-47cb-b528-7534d99126a4-data-volume\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.218211 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218104 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq6gl\" (UniqueName: \"kubernetes.io/projected/8315b63f-0468-47cb-b528-7534d99126a4-kube-api-access-sq6gl\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.218211 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218109 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0326059a-7ea7-44ae-bc43-7619240760e7-ca-trust-extracted\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.218211 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218134 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-tmp-dir\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.218211 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218194 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-bound-sa-token\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.218395 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218229 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/8315b63f-0468-47cb-b528-7534d99126a4-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.218395 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218269 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jhxzz\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-kube-api-access-jhxzz\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.218551 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.218531 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0326059a-7ea7-44ae-bc43-7619240760e7-trusted-ca\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.221550 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.221526 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0326059a-7ea7-44ae-bc43-7619240760e7-installation-pull-secrets\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.221642 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.221524 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/0326059a-7ea7-44ae-bc43-7619240760e7-image-registry-private-configuration\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.221642 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.221620 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-registry-tls\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.229190 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.229170 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhxzz\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-kube-api-access-jhxzz\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.229319 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.229304 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0326059a-7ea7-44ae-bc43-7619240760e7-registry-certificates\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.230375 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.230354 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0326059a-7ea7-44ae-bc43-7619240760e7-bound-sa-token\") pod \"image-registry-f6646996-g48x4\" (UID: \"0326059a-7ea7-44ae-bc43-7619240760e7\") " pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.318959 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.318882 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-metrics-tls\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.318959 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.318910 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8q7fq\" (UniqueName: \"kubernetes.io/projected/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-kube-api-access-8q7fq\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.319098 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319058 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/8315b63f-0468-47cb-b528-7534d99126a4-data-volume\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.319098 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319083 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sq6gl\" (UniqueName: \"kubernetes.io/projected/8315b63f-0468-47cb-b528-7534d99126a4-kube-api-access-sq6gl\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.319205 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319102 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-tmp-dir\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.319205 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319145 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/8315b63f-0468-47cb-b528-7534d99126a4-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.319205 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319179 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-47l6r\" (UniqueName: \"kubernetes.io/projected/629aaa70-005e-4af9-9ce5-ca15d667f5e7-kube-api-access-47l6r\") pod \"ingress-canary-b9bkd\" (UID: \"629aaa70-005e-4af9-9ce5-ca15d667f5e7\") " pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.319362 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319207 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/629aaa70-005e-4af9-9ce5-ca15d667f5e7-cert\") pod \"ingress-canary-b9bkd\" (UID: \"629aaa70-005e-4af9-9ce5-ca15d667f5e7\") " pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.319362 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319230 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-config-volume\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.319362 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319296 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/8315b63f-0468-47cb-b528-7534d99126a4-crio-socket\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.319362 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319347 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/8315b63f-0468-47cb-b528-7534d99126a4-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.319702 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319384 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/8315b63f-0468-47cb-b528-7534d99126a4-data-volume\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.319702 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319414 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/8315b63f-0468-47cb-b528-7534d99126a4-crio-socket\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.319702 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319692 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-tmp-dir\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.319859 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.319803 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/8315b63f-0468-47cb-b528-7534d99126a4-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.320134 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.320104 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-config-volume\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.321145 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.321124 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-metrics-tls\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.321476 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.321461 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/629aaa70-005e-4af9-9ce5-ca15d667f5e7-cert\") pod \"ingress-canary-b9bkd\" (UID: \"629aaa70-005e-4af9-9ce5-ca15d667f5e7\") " pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.321566 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.321552 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/8315b63f-0468-47cb-b528-7534d99126a4-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.329323 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.329305 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q7fq\" (UniqueName: \"kubernetes.io/projected/e39bb5e3-7f87-49f6-8be9-68b1fefc9b02-kube-api-access-8q7fq\") pod \"dns-default-57ft4\" (UID: \"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02\") " pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.329464 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.329444 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-47l6r\" (UniqueName: \"kubernetes.io/projected/629aaa70-005e-4af9-9ce5-ca15d667f5e7-kube-api-access-47l6r\") pod \"ingress-canary-b9bkd\" (UID: \"629aaa70-005e-4af9-9ce5-ca15d667f5e7\") " pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.330390 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.330372 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq6gl\" (UniqueName: \"kubernetes.io/projected/8315b63f-0468-47cb-b528-7534d99126a4-kube-api-access-sq6gl\") pod \"insights-runtime-extractor-sh9jw\" (UID: \"8315b63f-0468-47cb-b528-7534d99126a4\") " pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.403011 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.402995 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.445843 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.445818 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:34.469524 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.469498 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b9bkd" Apr 22 18:21:34.485996 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.485663 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-sh9jw" Apr 22 18:21:34.613136 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.613081 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-f6646996-g48x4"] Apr 22 18:21:34.617387 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:34.617359 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0326059a_7ea7_44ae_bc43_7619240760e7.slice/crio-65a2b2ac790024697ae7e7b0945465b3285003e126052dbb8ed98da501288fbb WatchSource:0}: Error finding container 65a2b2ac790024697ae7e7b0945465b3285003e126052dbb8ed98da501288fbb: Status 404 returned error can't find the container with id 65a2b2ac790024697ae7e7b0945465b3285003e126052dbb8ed98da501288fbb Apr 22 18:21:34.626718 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.626694 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-57ft4"] Apr 22 18:21:34.629391 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:34.629366 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode39bb5e3_7f87_49f6_8be9_68b1fefc9b02.slice/crio-ed24130ee035dee0c01324e2258fd3207c5f56d57348d260166eafea7909ca45 WatchSource:0}: Error finding container ed24130ee035dee0c01324e2258fd3207c5f56d57348d260166eafea7909ca45: Status 404 returned error can't find the container with id ed24130ee035dee0c01324e2258fd3207c5f56d57348d260166eafea7909ca45 Apr 22 18:21:34.642078 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.642055 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b9bkd"] Apr 22 18:21:34.644809 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:34.644786 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod629aaa70_005e_4af9_9ce5_ca15d667f5e7.slice/crio-e192d97cbd73f2c4d190d555f3590fedae5ca34a025699ee1f496cf8a85bab2e WatchSource:0}: Error finding container e192d97cbd73f2c4d190d555f3590fedae5ca34a025699ee1f496cf8a85bab2e: Status 404 returned error can't find the container with id e192d97cbd73f2c4d190d555f3590fedae5ca34a025699ee1f496cf8a85bab2e Apr 22 18:21:34.661521 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.661502 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-sh9jw"] Apr 22 18:21:34.664881 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:34.664861 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8315b63f_0468_47cb_b528_7534d99126a4.slice/crio-b28b31f2f53e3e430b45247d357667475dd96200b24878ac08ccfc02bcaccdb7 WatchSource:0}: Error finding container b28b31f2f53e3e430b45247d357667475dd96200b24878ac08ccfc02bcaccdb7: Status 404 returned error can't find the container with id b28b31f2f53e3e430b45247d357667475dd96200b24878ac08ccfc02bcaccdb7 Apr 22 18:21:34.953620 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.953562 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-57ft4" event={"ID":"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02","Type":"ContainerStarted","Data":"ed24130ee035dee0c01324e2258fd3207c5f56d57348d260166eafea7909ca45"} Apr 22 18:21:34.954811 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.954779 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-sh9jw" event={"ID":"8315b63f-0468-47cb-b528-7534d99126a4","Type":"ContainerStarted","Data":"272984e040126170e41650dc16b10f3f796d2f662d34b4932ff6b859440fb620"} Apr 22 18:21:34.954811 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.954808 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-sh9jw" event={"ID":"8315b63f-0468-47cb-b528-7534d99126a4","Type":"ContainerStarted","Data":"b28b31f2f53e3e430b45247d357667475dd96200b24878ac08ccfc02bcaccdb7"} Apr 22 18:21:34.955745 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.955721 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b9bkd" event={"ID":"629aaa70-005e-4af9-9ce5-ca15d667f5e7","Type":"ContainerStarted","Data":"e192d97cbd73f2c4d190d555f3590fedae5ca34a025699ee1f496cf8a85bab2e"} Apr 22 18:21:34.956973 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.956950 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-f6646996-g48x4" event={"ID":"0326059a-7ea7-44ae-bc43-7619240760e7","Type":"ContainerStarted","Data":"1f2f1aa8136ad8fbb27831f422e3938be66449360d3e2abcddf00e30fa7c3071"} Apr 22 18:21:34.957058 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.956979 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-f6646996-g48x4" event={"ID":"0326059a-7ea7-44ae-bc43-7619240760e7","Type":"ContainerStarted","Data":"65a2b2ac790024697ae7e7b0945465b3285003e126052dbb8ed98da501288fbb"} Apr 22 18:21:34.957105 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.957088 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:34.979326 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:34.979277 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-f6646996-g48x4" podStartSLOduration=0.979261345 podStartE2EDuration="979.261345ms" podCreationTimestamp="2026-04-22 18:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 18:21:34.97908781 +0000 UTC m=+51.783151188" watchObservedRunningTime="2026-04-22 18:21:34.979261345 +0000 UTC m=+51.783324709" Apr 22 18:21:35.739151 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.739116 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:35.739331 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.739244 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:35.739470 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.739116 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:35.743120 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.743099 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 22 18:21:35.743221 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.743159 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 22 18:21:35.743286 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.743159 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-bn55n\"" Apr 22 18:21:35.743286 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.743251 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 22 18:21:35.744253 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.744230 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-zkrzn\"" Apr 22 18:21:35.744881 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:35.744559 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 22 18:21:37.968335 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:37.968302 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b9bkd" event={"ID":"629aaa70-005e-4af9-9ce5-ca15d667f5e7","Type":"ContainerStarted","Data":"b8bfefd6fce6ee56efccd136814bf8f83fb12efec7538b01086997fd738cd415"} Apr 22 18:21:37.969865 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:37.969839 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-57ft4" event={"ID":"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02","Type":"ContainerStarted","Data":"7e18111dcba85c12debef89c198e8403e90e55e83808c70fed38a8a6e135a174"} Apr 22 18:21:37.969962 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:37.969874 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-57ft4" event={"ID":"e39bb5e3-7f87-49f6-8be9-68b1fefc9b02","Type":"ContainerStarted","Data":"cce528c12b692216fb2e56373ed7b2180ac5455b2aa1a024ed4846e0a041be02"} Apr 22 18:21:37.969962 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:37.969891 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:37.971201 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:37.971184 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-sh9jw" event={"ID":"8315b63f-0468-47cb-b528-7534d99126a4","Type":"ContainerStarted","Data":"f34e0de92424425547023c8cdb8284c3fe31581c9b17d2ac7c1a22c146bc467c"} Apr 22 18:21:37.984181 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:37.984142 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b9bkd" podStartSLOduration=1.489017067 podStartE2EDuration="3.984130936s" podCreationTimestamp="2026-04-22 18:21:34 +0000 UTC" firstStartedPulling="2026-04-22 18:21:34.64648957 +0000 UTC m=+51.450552917" lastFinishedPulling="2026-04-22 18:21:37.14160343 +0000 UTC m=+53.945666786" observedRunningTime="2026-04-22 18:21:37.98363859 +0000 UTC m=+54.787701956" watchObservedRunningTime="2026-04-22 18:21:37.984130936 +0000 UTC m=+54.788194293" Apr 22 18:21:38.000144 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:38.000110 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-57ft4" podStartSLOduration=1.493741407 podStartE2EDuration="4.00010177s" podCreationTimestamp="2026-04-22 18:21:34 +0000 UTC" firstStartedPulling="2026-04-22 18:21:34.630964426 +0000 UTC m=+51.435027770" lastFinishedPulling="2026-04-22 18:21:37.137324784 +0000 UTC m=+53.941388133" observedRunningTime="2026-04-22 18:21:37.999603445 +0000 UTC m=+54.803666811" watchObservedRunningTime="2026-04-22 18:21:38.00010177 +0000 UTC m=+54.804165134" Apr 22 18:21:40.982126 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:40.982087 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-sh9jw" event={"ID":"8315b63f-0468-47cb-b528-7534d99126a4","Type":"ContainerStarted","Data":"42d10a13b53bcaad421a711beb2bdf69abc71564a2f0c4c9866726586bc87d5f"} Apr 22 18:21:41.006424 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:41.006383 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-sh9jw" podStartSLOduration=1.5142136229999998 podStartE2EDuration="7.006371367s" podCreationTimestamp="2026-04-22 18:21:34 +0000 UTC" firstStartedPulling="2026-04-22 18:21:34.744007186 +0000 UTC m=+51.548070533" lastFinishedPulling="2026-04-22 18:21:40.236164931 +0000 UTC m=+57.040228277" observedRunningTime="2026-04-22 18:21:41.005816233 +0000 UTC m=+57.809879593" watchObservedRunningTime="2026-04-22 18:21:41.006371367 +0000 UTC m=+57.810434766" Apr 22 18:21:41.915147 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:41.915117 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-44fqr" Apr 22 18:21:47.977863 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:47.977827 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-57ft4" Apr 22 18:21:49.425063 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.425023 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:49.427826 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.427807 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 22 18:21:49.438094 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.438067 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5ac120e-7d32-45b8-a5d6-57aa386e732f-metrics-certs\") pod \"network-metrics-daemon-xl5pv\" (UID: \"b5ac120e-7d32-45b8-a5d6-57aa386e732f\") " pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:49.525889 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.525860 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:49.528812 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.528798 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 22 18:21:49.539062 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.539043 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 22 18:21:49.548502 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.548486 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbnfq\" (UniqueName: \"kubernetes.io/projected/e4946d3e-bc1b-4f87-8fb8-db35971cf837-kube-api-access-gbnfq\") pod \"network-check-target-47xwd\" (UID: \"e4946d3e-bc1b-4f87-8fb8-db35971cf837\") " pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:49.558934 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.558907 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-zkrzn\"" Apr 22 18:21:49.565095 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.565082 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:49.570895 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.570879 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-bn55n\"" Apr 22 18:21:49.578297 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.578282 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xl5pv" Apr 22 18:21:49.685958 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.685877 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-47xwd"] Apr 22 18:21:49.689368 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:49.689321 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4946d3e_bc1b_4f87_8fb8_db35971cf837.slice/crio-7a1457323b7bc84632109e125305e8d7f63fc987d0bab6ad81607c61ac670da7 WatchSource:0}: Error finding container 7a1457323b7bc84632109e125305e8d7f63fc987d0bab6ad81607c61ac670da7: Status 404 returned error can't find the container with id 7a1457323b7bc84632109e125305e8d7f63fc987d0bab6ad81607c61ac670da7 Apr 22 18:21:49.704257 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:49.704231 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xl5pv"] Apr 22 18:21:49.707014 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:49.706994 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5ac120e_7d32_45b8_a5d6_57aa386e732f.slice/crio-c212e2c6aeecb7ac1d32ab68876eeb9de3f60474265c7529f76af02f0afb66d6 WatchSource:0}: Error finding container c212e2c6aeecb7ac1d32ab68876eeb9de3f60474265c7529f76af02f0afb66d6: Status 404 returned error can't find the container with id c212e2c6aeecb7ac1d32ab68876eeb9de3f60474265c7529f76af02f0afb66d6 Apr 22 18:21:50.006000 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:50.005963 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-47xwd" event={"ID":"e4946d3e-bc1b-4f87-8fb8-db35971cf837","Type":"ContainerStarted","Data":"7a1457323b7bc84632109e125305e8d7f63fc987d0bab6ad81607c61ac670da7"} Apr 22 18:21:50.006894 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:50.006871 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xl5pv" event={"ID":"b5ac120e-7d32-45b8-a5d6-57aa386e732f","Type":"ContainerStarted","Data":"c212e2c6aeecb7ac1d32ab68876eeb9de3f60474265c7529f76af02f0afb66d6"} Apr 22 18:21:51.011819 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:51.011770 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xl5pv" event={"ID":"b5ac120e-7d32-45b8-a5d6-57aa386e732f","Type":"ContainerStarted","Data":"580f911b53391c630bcad5b1bdb830e50e242a684db17d5f40a89ec1d22c2049"} Apr 22 18:21:52.016982 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.016945 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xl5pv" event={"ID":"b5ac120e-7d32-45b8-a5d6-57aa386e732f","Type":"ContainerStarted","Data":"a7fcf5fb14476236fad428a788991b8832567c4a78c0e95cd981a3a8e53d78b2"} Apr 22 18:21:52.040086 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.040032 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xl5pv" podStartSLOduration=67.963581205 podStartE2EDuration="1m9.040015262s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:21:49.708643525 +0000 UTC m=+66.512706872" lastFinishedPulling="2026-04-22 18:21:50.785077572 +0000 UTC m=+67.589140929" observedRunningTime="2026-04-22 18:21:52.03918481 +0000 UTC m=+68.843248174" watchObservedRunningTime="2026-04-22 18:21:52.040015262 +0000 UTC m=+68.844078628" Apr 22 18:21:52.094531 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.094505 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-njfv5"] Apr 22 18:21:52.097672 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.097642 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.100331 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.100301 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 22 18:21:52.100734 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.100627 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 22 18:21:52.100734 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.100665 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 22 18:21:52.100734 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.100714 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 22 18:21:52.100958 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.100742 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-ghp22\"" Apr 22 18:21:52.100958 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.100750 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 22 18:21:52.100958 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.100831 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 22 18:21:52.144337 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144310 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a33be29f-8691-4d49-8d7c-0b6d40636abc-metrics-client-ca\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144437 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144344 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-275st\" (UniqueName: \"kubernetes.io/projected/a33be29f-8691-4d49-8d7c-0b6d40636abc-kube-api-access-275st\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144437 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144379 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-accelerators-collector-config\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144516 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-sys\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144588 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144545 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-tls\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144588 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144567 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-textfile\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144685 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144665 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-wtmp\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144727 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144695 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.144727 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.144714 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-root\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245192 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245160 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a33be29f-8691-4d49-8d7c-0b6d40636abc-metrics-client-ca\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245342 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245252 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-275st\" (UniqueName: \"kubernetes.io/projected/a33be29f-8691-4d49-8d7c-0b6d40636abc-kube-api-access-275st\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245342 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245318 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-accelerators-collector-config\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245453 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245384 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-sys\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245453 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245411 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-tls\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245453 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245441 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-textfile\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245593 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245482 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-wtmp\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245593 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245496 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-sys\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245593 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245510 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245817 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245796 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a33be29f-8691-4d49-8d7c-0b6d40636abc-metrics-client-ca\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.245963 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:52.245945 2581 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Apr 22 18:21:52.246041 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245973 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-accelerators-collector-config\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.246041 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.245998 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-root\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.246041 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:52.246021 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-tls podName:a33be29f-8691-4d49-8d7c-0b6d40636abc nodeName:}" failed. No retries permitted until 2026-04-22 18:21:52.746001962 +0000 UTC m=+69.550065307 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-tls") pod "node-exporter-njfv5" (UID: "a33be29f-8691-4d49-8d7c-0b6d40636abc") : secret "node-exporter-tls" not found Apr 22 18:21:52.246178 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.246057 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-root\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.246178 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.246081 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-textfile\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.246258 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.246168 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-wtmp\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.248138 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.248118 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.253988 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.253966 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-275st\" (UniqueName: \"kubernetes.io/projected/a33be29f-8691-4d49-8d7c-0b6d40636abc-kube-api-access-275st\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.749212 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.749187 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-tls\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:52.751372 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:52.751347 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a33be29f-8691-4d49-8d7c-0b6d40636abc-node-exporter-tls\") pod \"node-exporter-njfv5\" (UID: \"a33be29f-8691-4d49-8d7c-0b6d40636abc\") " pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:53.008256 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.008192 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-njfv5" Apr 22 18:21:53.016231 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:53.016211 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda33be29f_8691_4d49_8d7c_0b6d40636abc.slice/crio-1dbb9e54a27642200eacfd1516aafb6b5de0e346a521ce93edd0de478f4c0a3a WatchSource:0}: Error finding container 1dbb9e54a27642200eacfd1516aafb6b5de0e346a521ce93edd0de478f4c0a3a: Status 404 returned error can't find the container with id 1dbb9e54a27642200eacfd1516aafb6b5de0e346a521ce93edd0de478f4c0a3a Apr 22 18:21:53.020214 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.020191 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-njfv5" event={"ID":"a33be29f-8691-4d49-8d7c-0b6d40636abc","Type":"ContainerStarted","Data":"1dbb9e54a27642200eacfd1516aafb6b5de0e346a521ce93edd0de478f4c0a3a"} Apr 22 18:21:53.021293 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.021273 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-47xwd" event={"ID":"e4946d3e-bc1b-4f87-8fb8-db35971cf837","Type":"ContainerStarted","Data":"1731e8b16fde2f88dcb388ff366c9d6f4f62e13027fd04966e9caea886015d64"} Apr 22 18:21:53.021429 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.021416 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:21:53.046334 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.046294 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-47xwd" podStartSLOduration=67.321714858 podStartE2EDuration="1m10.046281985s" podCreationTimestamp="2026-04-22 18:20:43 +0000 UTC" firstStartedPulling="2026-04-22 18:21:49.691638605 +0000 UTC m=+66.495701948" lastFinishedPulling="2026-04-22 18:21:52.416205728 +0000 UTC m=+69.220269075" observedRunningTime="2026-04-22 18:21:53.045647447 +0000 UTC m=+69.849710812" watchObservedRunningTime="2026-04-22 18:21:53.046281985 +0000 UTC m=+69.850345378" Apr 22 18:21:53.156522 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.156501 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:21:53.161037 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.161021 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.164912 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.164895 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Apr 22 18:21:53.165172 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.165160 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Apr 22 18:21:53.165511 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.165499 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Apr 22 18:21:53.165759 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.165736 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Apr 22 18:21:53.165881 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.165741 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Apr 22 18:21:53.165881 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.165774 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Apr 22 18:21:53.165881 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.165787 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Apr 22 18:21:53.165881 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.165784 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Apr 22 18:21:53.166187 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.166171 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-2hrm4\"" Apr 22 18:21:53.167236 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.166958 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Apr 22 18:21:53.178541 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.178523 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:21:53.253316 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253295 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253437 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253326 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9679r\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-kube-api-access-9679r\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253437 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253345 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253437 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253407 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-tls-assets\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253437 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253434 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-web-config\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253463 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253488 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253515 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253536 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253558 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253584 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-volume\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253605 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-out\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.253637 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.253622 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.354582 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354525 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.354582 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354552 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-volume\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.354582 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354568 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-out\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.354779 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354586 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.354869 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354835 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.354950 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354878 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9679r\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-kube-api-access-9679r\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.354950 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354879 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355058 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.354907 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355058 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:53.354983 2581 secret.go:189] Couldn't get secret openshift-monitoring/alertmanager-main-tls: secret "alertmanager-main-tls" not found Apr 22 18:21:53.355058 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.355017 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-tls-assets\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355058 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:53.355046 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls podName:4c77cb4e-4729-4bfa-ac40-e4e0837818f3 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:53.855027807 +0000 UTC m=+70.659091328 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-alertmanager-main-tls" (UniqueName: "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls") pod "alertmanager-main-0" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3") : secret "alertmanager-main-tls" not found Apr 22 18:21:53.355323 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.355073 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-web-config\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355323 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.355105 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355323 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.355153 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355323 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.355183 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355323 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.355219 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.355560 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:53.355367 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle podName:4c77cb4e-4729-4bfa-ac40-e4e0837818f3 nodeName:}" failed. No retries permitted until 2026-04-22 18:21:53.85534738 +0000 UTC m=+70.659410739 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "alertmanager-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle") pod "alertmanager-main-0" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3") : configmap references non-existent config key: ca-bundle.crt Apr 22 18:21:53.355560 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.355398 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.357581 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.357511 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-out\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.357732 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.357684 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.357969 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.357951 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-tls-assets\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.358294 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.358274 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-web-config\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.358352 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.358306 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.358693 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.358672 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-volume\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.358755 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.358690 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.358755 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.358692 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.367979 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.367959 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9679r\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-kube-api-access-9679r\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.858592 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.858483 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.858592 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.858527 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.859285 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.859254 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:53.860795 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:53.860775 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:54.025337 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.025301 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-njfv5" event={"ID":"a33be29f-8691-4d49-8d7c-0b6d40636abc","Type":"ContainerStarted","Data":"eb7f28eb812505198f87fee3630e90bb6ab06b749651dd86316ab801becf233e"} Apr 22 18:21:54.070026 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.070001 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:21:54.141968 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.141911 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-54f58b6d86-znvnl"] Apr 22 18:21:54.146841 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.146820 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.149386 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.149361 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy\"" Apr 22 18:21:54.149491 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.149444 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-web\"" Apr 22 18:21:54.149740 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.149701 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-metrics\"" Apr 22 18:21:54.149832 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.149792 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-rules\"" Apr 22 18:21:54.149984 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.149915 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-tls\"" Apr 22 18:21:54.150072 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.150052 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-dockercfg-g554b\"" Apr 22 18:21:54.150132 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.150087 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-grpc-tls-ai0ahusqhf6oi\"" Apr 22 18:21:54.158486 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.158462 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-54f58b6d86-znvnl"] Apr 22 18:21:54.194278 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.194259 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:21:54.197319 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:54.197286 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c77cb4e_4729_4bfa_ac40_e4e0837818f3.slice/crio-a39fb4fb56689c82ceed47b3ca569766f21802060dee2320678e96645aa577bd WatchSource:0}: Error finding container a39fb4fb56689c82ceed47b3ca569766f21802060dee2320678e96645aa577bd: Status 404 returned error can't find the container with id a39fb4fb56689c82ceed47b3ca569766f21802060dee2320678e96645aa577bd Apr 22 18:21:54.262075 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262050 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.262154 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262079 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/47c2b707-b606-473b-b666-0b19f7d15ecb-metrics-client-ca\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.262154 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262098 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.262154 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262126 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.262253 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262193 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.262253 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262231 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chktb\" (UniqueName: \"kubernetes.io/projected/47c2b707-b606-473b-b666-0b19f7d15ecb-kube-api-access-chktb\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.262253 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262247 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-grpc-tls\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.262346 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.262262 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-tls\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363000 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.362953 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363000 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.362994 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363117 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.363025 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363154 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.363135 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-chktb\" (UniqueName: \"kubernetes.io/projected/47c2b707-b606-473b-b666-0b19f7d15ecb-kube-api-access-chktb\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363199 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.363164 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-grpc-tls\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363199 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.363185 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-tls\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363293 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.363220 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.363293 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.363252 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/47c2b707-b606-473b-b666-0b19f7d15ecb-metrics-client-ca\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.364131 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.364108 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/47c2b707-b606-473b-b666-0b19f7d15ecb-metrics-client-ca\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.365669 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.365645 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.365959 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.365940 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.366064 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.366042 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.366143 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.366128 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-grpc-tls\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.366307 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.366287 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-tls\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.366343 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.366290 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/47c2b707-b606-473b-b666-0b19f7d15ecb-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.371473 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.371457 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-chktb\" (UniqueName: \"kubernetes.io/projected/47c2b707-b606-473b-b666-0b19f7d15ecb-kube-api-access-chktb\") pod \"thanos-querier-54f58b6d86-znvnl\" (UID: \"47c2b707-b606-473b-b666-0b19f7d15ecb\") " pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.406750 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.406725 2581 patch_prober.go:28] interesting pod/image-registry-f6646996-g48x4 container/registry namespace/openshift-image-registry: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 22 18:21:54.406835 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.406763 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-f6646996-g48x4" podUID="0326059a-7ea7-44ae-bc43-7619240760e7" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 22 18:21:54.458818 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.458795 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:21:54.463574 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.463550 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:54.466485 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.466443 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 22 18:21:54.475799 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.475781 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f0d1648a-0bbf-4049-b68e-82046b4578f0-original-pull-secret\") pod \"global-pull-secret-syncer-x5xw7\" (UID: \"f0d1648a-0bbf-4049-b68e-82046b4578f0\") " pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:54.602655 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.602622 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-54f58b6d86-znvnl"] Apr 22 18:21:54.606424 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:54.606399 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47c2b707_b606_473b_b666_0b19f7d15ecb.slice/crio-e50c55e9b849e90577a01a92d9a246ee8601e177e685962854e6c437be66a562 WatchSource:0}: Error finding container e50c55e9b849e90577a01a92d9a246ee8601e177e685962854e6c437be66a562: Status 404 returned error can't find the container with id e50c55e9b849e90577a01a92d9a246ee8601e177e685962854e6c437be66a562 Apr 22 18:21:54.661537 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.661489 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-x5xw7" Apr 22 18:21:54.773950 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:54.773900 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-x5xw7"] Apr 22 18:21:54.776929 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:54.776889 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0d1648a_0bbf_4049_b68e_82046b4578f0.slice/crio-b4c103895de4331a2679d2485e05a6873ce56badcac03190b1c037ea7b3eff16 WatchSource:0}: Error finding container b4c103895de4331a2679d2485e05a6873ce56badcac03190b1c037ea7b3eff16: Status 404 returned error can't find the container with id b4c103895de4331a2679d2485e05a6873ce56badcac03190b1c037ea7b3eff16 Apr 22 18:21:55.034785 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:55.034751 2581 generic.go:358] "Generic (PLEG): container finished" podID="a33be29f-8691-4d49-8d7c-0b6d40636abc" containerID="eb7f28eb812505198f87fee3630e90bb6ab06b749651dd86316ab801becf233e" exitCode=0 Apr 22 18:21:55.035205 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:55.034838 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-njfv5" event={"ID":"a33be29f-8691-4d49-8d7c-0b6d40636abc","Type":"ContainerDied","Data":"eb7f28eb812505198f87fee3630e90bb6ab06b749651dd86316ab801becf233e"} Apr 22 18:21:55.035977 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:55.035951 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-x5xw7" event={"ID":"f0d1648a-0bbf-4049-b68e-82046b4578f0","Type":"ContainerStarted","Data":"b4c103895de4331a2679d2485e05a6873ce56badcac03190b1c037ea7b3eff16"} Apr 22 18:21:55.036969 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:55.036946 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" event={"ID":"47c2b707-b606-473b-b666-0b19f7d15ecb","Type":"ContainerStarted","Data":"e50c55e9b849e90577a01a92d9a246ee8601e177e685962854e6c437be66a562"} Apr 22 18:21:55.037959 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:55.037934 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerStarted","Data":"a39fb4fb56689c82ceed47b3ca569766f21802060dee2320678e96645aa577bd"} Apr 22 18:21:55.966803 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:55.966732 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-f6646996-g48x4" Apr 22 18:21:56.043518 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.043478 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-njfv5" event={"ID":"a33be29f-8691-4d49-8d7c-0b6d40636abc","Type":"ContainerStarted","Data":"9d5e927fa42db46c599c018cfc04f4fd01d0ee423a976e30a83e9a8a4336f0c6"} Apr 22 18:21:56.043518 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.043520 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-njfv5" event={"ID":"a33be29f-8691-4d49-8d7c-0b6d40636abc","Type":"ContainerStarted","Data":"6a27c19f2a7bc9dac3c3bf2b3c07b6697cc00222e366af2ae36052adf9883e17"} Apr 22 18:21:56.045991 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.045902 2581 generic.go:358] "Generic (PLEG): container finished" podID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerID="1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416" exitCode=0 Apr 22 18:21:56.045991 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.045962 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416"} Apr 22 18:21:56.067100 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.067056 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-njfv5" podStartSLOduration=3.183772723 podStartE2EDuration="4.067043374s" podCreationTimestamp="2026-04-22 18:21:52 +0000 UTC" firstStartedPulling="2026-04-22 18:21:53.018063199 +0000 UTC m=+69.822126544" lastFinishedPulling="2026-04-22 18:21:53.901333851 +0000 UTC m=+70.705397195" observedRunningTime="2026-04-22 18:21:56.065773085 +0000 UTC m=+72.869836450" watchObservedRunningTime="2026-04-22 18:21:56.067043374 +0000 UTC m=+72.871106731" Apr 22 18:21:56.868318 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.868234 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj"] Apr 22 18:21:56.871790 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.871770 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:21:56.878833 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.878592 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"monitoring-plugin-cert\"" Apr 22 18:21:56.878833 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.878600 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"default-dockercfg-vhtwm\"" Apr 22 18:21:56.885301 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.885277 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj"] Apr 22 18:21:56.983757 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:56.983720 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-ccnhj\" (UID: \"3d87f815-f0b3-4790-bf18-63a31165c49a\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:21:57.055008 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.054974 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" event={"ID":"47c2b707-b606-473b-b666-0b19f7d15ecb","Type":"ContainerStarted","Data":"e1a1b17e23f9f9db9f36f36d26ffa25cc466658b220a030f38b5f6f97336a6f2"} Apr 22 18:21:57.055381 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.055017 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" event={"ID":"47c2b707-b606-473b-b666-0b19f7d15ecb","Type":"ContainerStarted","Data":"f604952c3f5cc529802fad794c46fefffc1b277bbfad649a8322a07a2ce8af35"} Apr 22 18:21:57.055381 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.055031 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" event={"ID":"47c2b707-b606-473b-b666-0b19f7d15ecb","Type":"ContainerStarted","Data":"e1b25dbbd7e339c7299d8d77c7d8de87c727cd48dc565ed763f5f5421310cbf3"} Apr 22 18:21:57.084660 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.084634 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-ccnhj\" (UID: \"3d87f815-f0b3-4790-bf18-63a31165c49a\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:21:57.084795 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:57.084781 2581 secret.go:189] Couldn't get secret openshift-monitoring/monitoring-plugin-cert: secret "monitoring-plugin-cert" not found Apr 22 18:21:57.084855 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:57.084849 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert podName:3d87f815-f0b3-4790-bf18-63a31165c49a nodeName:}" failed. No retries permitted until 2026-04-22 18:21:57.584834291 +0000 UTC m=+74.388897638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "monitoring-plugin-cert" (UniqueName: "kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert") pod "monitoring-plugin-7dccd58f55-ccnhj" (UID: "3d87f815-f0b3-4790-bf18-63a31165c49a") : secret "monitoring-plugin-cert" not found Apr 22 18:21:57.202176 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.202095 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/telemeter-client-6fc869648c-gt45s"] Apr 22 18:21:57.205364 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.205344 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.208371 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.208198 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemeter-client-serving-certs-ca-bundle\"" Apr 22 18:21:57.208371 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.208212 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-kube-rbac-proxy-config\"" Apr 22 18:21:57.208371 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.208199 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-dockercfg-t59gz\"" Apr 22 18:21:57.208371 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.208295 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-tls\"" Apr 22 18:21:57.208371 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.208315 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"federate-client-certs\"" Apr 22 18:21:57.208700 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.208201 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client\"" Apr 22 18:21:57.214717 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.214696 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemeter-trusted-ca-bundle-8i12ta5c71j38\"" Apr 22 18:21:57.221260 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.221232 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-6fc869648c-gt45s"] Apr 22 18:21:57.286987 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.286953 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-metrics-client-ca\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.287131 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.286997 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.287131 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.287026 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-serving-certs-ca-bundle\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.287242 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.287122 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-telemeter-client-tls\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.287242 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.287160 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-telemeter-trusted-ca-bundle\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.287343 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.287250 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnm2f\" (UniqueName: \"kubernetes.io/projected/fd1fd636-f128-44d6-9294-d1aad120ef52-kube-api-access-dnm2f\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.287343 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.287268 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-secret-telemeter-client\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.287343 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.287287 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-federate-client-tls\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388330 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388297 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-telemeter-client-tls\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388330 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388334 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-telemeter-trusted-ca-bundle\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388399 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnm2f\" (UniqueName: \"kubernetes.io/projected/fd1fd636-f128-44d6-9294-d1aad120ef52-kube-api-access-dnm2f\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388425 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-secret-telemeter-client\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388446 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-federate-client-tls\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388468 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-metrics-client-ca\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388502 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.388543 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.388529 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-serving-certs-ca-bundle\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.389401 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.389353 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-telemeter-trusted-ca-bundle\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.389893 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.389872 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-serving-certs-ca-bundle\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.389991 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.389877 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fd1fd636-f128-44d6-9294-d1aad120ef52-metrics-client-ca\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.391259 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.391216 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-secret-telemeter-client\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.391893 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.391483 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-telemeter-client-tls\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.391893 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.391866 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-federate-client-tls\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.392394 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.392373 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fd1fd636-f128-44d6-9294-d1aad120ef52-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.397264 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.397244 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnm2f\" (UniqueName: \"kubernetes.io/projected/fd1fd636-f128-44d6-9294-d1aad120ef52-kube-api-access-dnm2f\") pod \"telemeter-client-6fc869648c-gt45s\" (UID: \"fd1fd636-f128-44d6-9294-d1aad120ef52\") " pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.516770 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.516742 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" Apr 22 18:21:57.590240 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:57.590212 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-ccnhj\" (UID: \"3d87f815-f0b3-4790-bf18-63a31165c49a\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:21:57.590375 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:57.590359 2581 secret.go:189] Couldn't get secret openshift-monitoring/monitoring-plugin-cert: secret "monitoring-plugin-cert" not found Apr 22 18:21:57.590438 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:21:57.590429 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert podName:3d87f815-f0b3-4790-bf18-63a31165c49a nodeName:}" failed. No retries permitted until 2026-04-22 18:21:58.590412727 +0000 UTC m=+75.394476082 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "monitoring-plugin-cert" (UniqueName: "kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert") pod "monitoring-plugin-7dccd58f55-ccnhj" (UID: "3d87f815-f0b3-4790-bf18-63a31165c49a") : secret "monitoring-plugin-cert" not found Apr 22 18:21:58.598789 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:58.598743 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-ccnhj\" (UID: \"3d87f815-f0b3-4790-bf18-63a31165c49a\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:21:58.602837 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:58.602811 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/3d87f815-f0b3-4790-bf18-63a31165c49a-monitoring-plugin-cert\") pod \"monitoring-plugin-7dccd58f55-ccnhj\" (UID: \"3d87f815-f0b3-4790-bf18-63a31165c49a\") " pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:21:58.682267 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:58.682242 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:21:59.299724 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:59.299696 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-6fc869648c-gt45s"] Apr 22 18:21:59.303511 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:59.303479 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd1fd636_f128_44d6_9294_d1aad120ef52.slice/crio-95cdcf101def96ca74514388014cae25253b8e802332dafe652ff54c9499fedb WatchSource:0}: Error finding container 95cdcf101def96ca74514388014cae25253b8e802332dafe652ff54c9499fedb: Status 404 returned error can't find the container with id 95cdcf101def96ca74514388014cae25253b8e802332dafe652ff54c9499fedb Apr 22 18:21:59.318570 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:21:59.318541 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj"] Apr 22 18:21:59.324939 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:21:59.324892 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d87f815_f0b3_4790_bf18_63a31165c49a.slice/crio-24d75928dff2eaed6a29c60db86f5ceadfa0fb6164758bd1222cc279000f1d26 WatchSource:0}: Error finding container 24d75928dff2eaed6a29c60db86f5ceadfa0fb6164758bd1222cc279000f1d26: Status 404 returned error can't find the container with id 24d75928dff2eaed6a29c60db86f5ceadfa0fb6164758bd1222cc279000f1d26 Apr 22 18:22:00.066492 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.066440 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-x5xw7" event={"ID":"f0d1648a-0bbf-4049-b68e-82046b4578f0","Type":"ContainerStarted","Data":"d6547603a8ae87eefbfafd115a01319b4e428344317837e79deab93f146b6fb1"} Apr 22 18:22:00.067590 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.067557 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" event={"ID":"fd1fd636-f128-44d6-9294-d1aad120ef52","Type":"ContainerStarted","Data":"95cdcf101def96ca74514388014cae25253b8e802332dafe652ff54c9499fedb"} Apr 22 18:22:00.070437 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.070413 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" event={"ID":"47c2b707-b606-473b-b666-0b19f7d15ecb","Type":"ContainerStarted","Data":"973aa78195ee7896a7ba3eec4b41c991e9d699a3716158d4953f38188f17be93"} Apr 22 18:22:00.070528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.070441 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" event={"ID":"47c2b707-b606-473b-b666-0b19f7d15ecb","Type":"ContainerStarted","Data":"264fc75bf2229843e013e50e0f27e05da2811f2188b7e1753f7213f4f259debf"} Apr 22 18:22:00.070528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.070451 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" event={"ID":"47c2b707-b606-473b-b666-0b19f7d15ecb","Type":"ContainerStarted","Data":"e192be4fe7fe56703ed5ca0d784bc6eccaf25c34147cc195d78d277240a63005"} Apr 22 18:22:00.070647 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.070631 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:22:00.071492 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.071472 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" event={"ID":"3d87f815-f0b3-4790-bf18-63a31165c49a","Type":"ContainerStarted","Data":"24d75928dff2eaed6a29c60db86f5ceadfa0fb6164758bd1222cc279000f1d26"} Apr 22 18:22:00.073944 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.073913 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerStarted","Data":"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7"} Apr 22 18:22:00.074021 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.073947 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerStarted","Data":"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1"} Apr 22 18:22:00.074021 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.073970 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerStarted","Data":"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563"} Apr 22 18:22:00.074021 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.073987 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerStarted","Data":"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044"} Apr 22 18:22:00.074021 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.073996 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerStarted","Data":"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98"} Apr 22 18:22:00.074021 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.074003 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerStarted","Data":"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5"} Apr 22 18:22:00.082367 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.082325 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-x5xw7" podStartSLOduration=65.717958752 podStartE2EDuration="1m10.082314094s" podCreationTimestamp="2026-04-22 18:20:50 +0000 UTC" firstStartedPulling="2026-04-22 18:21:54.778596769 +0000 UTC m=+71.582660112" lastFinishedPulling="2026-04-22 18:21:59.142952111 +0000 UTC m=+75.947015454" observedRunningTime="2026-04-22 18:22:00.081319768 +0000 UTC m=+76.885383134" watchObservedRunningTime="2026-04-22 18:22:00.082314094 +0000 UTC m=+76.886377458" Apr 22 18:22:00.109302 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.109258 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.164157634 podStartE2EDuration="7.109243463s" podCreationTimestamp="2026-04-22 18:21:53 +0000 UTC" firstStartedPulling="2026-04-22 18:21:54.199096271 +0000 UTC m=+71.003159618" lastFinishedPulling="2026-04-22 18:21:59.144182104 +0000 UTC m=+75.948245447" observedRunningTime="2026-04-22 18:22:00.107085177 +0000 UTC m=+76.911148554" watchObservedRunningTime="2026-04-22 18:22:00.109243463 +0000 UTC m=+76.913306829" Apr 22 18:22:00.129392 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:00.129347 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" podStartSLOduration=1.592865267 podStartE2EDuration="6.129332767s" podCreationTimestamp="2026-04-22 18:21:54 +0000 UTC" firstStartedPulling="2026-04-22 18:21:54.608231636 +0000 UTC m=+71.412294979" lastFinishedPulling="2026-04-22 18:21:59.14469912 +0000 UTC m=+75.948762479" observedRunningTime="2026-04-22 18:22:00.128355169 +0000 UTC m=+76.932418535" watchObservedRunningTime="2026-04-22 18:22:00.129332767 +0000 UTC m=+76.933396132" Apr 22 18:22:01.078165 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:01.078074 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" event={"ID":"3d87f815-f0b3-4790-bf18-63a31165c49a","Type":"ContainerStarted","Data":"93c6068dc96957e443cda67da0cb4cb063f8f7241e2f921708bc529e9497afad"} Apr 22 18:22:01.078547 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:01.078524 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:22:01.082862 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:01.082837 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" Apr 22 18:22:01.083324 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:01.083306 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-54f58b6d86-znvnl" Apr 22 18:22:01.097072 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:01.097037 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-7dccd58f55-ccnhj" podStartSLOduration=3.664298733 podStartE2EDuration="5.097025212s" podCreationTimestamp="2026-04-22 18:21:56 +0000 UTC" firstStartedPulling="2026-04-22 18:21:59.327299387 +0000 UTC m=+76.131362730" lastFinishedPulling="2026-04-22 18:22:00.760025866 +0000 UTC m=+77.564089209" observedRunningTime="2026-04-22 18:22:01.096113747 +0000 UTC m=+77.900177110" watchObservedRunningTime="2026-04-22 18:22:01.097025212 +0000 UTC m=+77.901088630" Apr 22 18:22:02.082988 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:02.082945 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" event={"ID":"fd1fd636-f128-44d6-9294-d1aad120ef52","Type":"ContainerStarted","Data":"cae92f04861d1b223df23dd82bfb0f2afffb892694a82b988f6f9ade7e28ba13"} Apr 22 18:22:02.083333 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:02.082996 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" event={"ID":"fd1fd636-f128-44d6-9294-d1aad120ef52","Type":"ContainerStarted","Data":"9e8391a125d823a1a496ca639f2bc7baf4f3ea6877588e21c50272d8550d4f26"} Apr 22 18:22:02.083333 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:02.083012 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" event={"ID":"fd1fd636-f128-44d6-9294-d1aad120ef52","Type":"ContainerStarted","Data":"4cf66a2ed347d6bb616157e5bbc289ee6a3c49dda9327bc25efa82c46a967068"} Apr 22 18:22:02.108412 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:02.108368 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/telemeter-client-6fc869648c-gt45s" podStartSLOduration=2.7341094679999998 podStartE2EDuration="5.10835547s" podCreationTimestamp="2026-04-22 18:21:57 +0000 UTC" firstStartedPulling="2026-04-22 18:21:59.307119466 +0000 UTC m=+76.111182811" lastFinishedPulling="2026-04-22 18:22:01.681365462 +0000 UTC m=+78.485428813" observedRunningTime="2026-04-22 18:22:02.107132541 +0000 UTC m=+78.911195907" watchObservedRunningTime="2026-04-22 18:22:02.10835547 +0000 UTC m=+78.912418834" Apr 22 18:22:04.244307 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.244275 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-57cf77bdb6-2vpwp"] Apr 22 18:22:04.248661 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.248643 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.251363 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.251343 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 22 18:22:04.251463 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.251448 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 22 18:22:04.251509 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.251473 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 22 18:22:04.252509 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.252494 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 22 18:22:04.252722 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.252699 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 22 18:22:04.252722 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.252730 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 22 18:22:04.252878 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.252797 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 22 18:22:04.253013 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.252999 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-jj6h5\"" Apr 22 18:22:04.257099 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.257079 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-57cf77bdb6-2vpwp"] Apr 22 18:22:04.257265 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.257249 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 22 18:22:04.345389 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.345363 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-config\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.345483 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.345390 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-oauth-serving-cert\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.345483 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.345411 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnjpq\" (UniqueName: \"kubernetes.io/projected/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-kube-api-access-jnjpq\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.345483 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.345430 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-trusted-ca-bundle\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.345483 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.345449 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-serving-cert\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.345666 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.345526 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-service-ca\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.345666 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.345593 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-oauth-config\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.445862 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.445839 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-service-ca\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.445972 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.445870 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-oauth-config\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.445972 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.445900 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-config\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.445972 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.445938 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-oauth-serving-cert\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.445972 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.445969 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jnjpq\" (UniqueName: \"kubernetes.io/projected/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-kube-api-access-jnjpq\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.446143 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.446010 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-trusted-ca-bundle\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.446143 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.446039 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-serving-cert\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.446559 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.446533 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-service-ca\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.446654 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.446631 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-oauth-serving-cert\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.446860 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.446841 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-config\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.447146 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.447127 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-trusted-ca-bundle\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.448527 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.448505 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-oauth-config\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.448595 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.448581 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-serving-cert\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.454340 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.454319 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnjpq\" (UniqueName: \"kubernetes.io/projected/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-kube-api-access-jnjpq\") pod \"console-57cf77bdb6-2vpwp\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.558352 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.558325 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:04.674984 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:04.674901 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-57cf77bdb6-2vpwp"] Apr 22 18:22:04.677329 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:22:04.677293 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf26a7e6a_9a08_4fe7_98dd_ddbe71e4fb98.slice/crio-a61f79e9a6bb7cd95ab84276ad6485d41995ed1040e5ccc126c001a1ecdf78f8 WatchSource:0}: Error finding container a61f79e9a6bb7cd95ab84276ad6485d41995ed1040e5ccc126c001a1ecdf78f8: Status 404 returned error can't find the container with id a61f79e9a6bb7cd95ab84276ad6485d41995ed1040e5ccc126c001a1ecdf78f8 Apr 22 18:22:05.092884 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:05.092851 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-57cf77bdb6-2vpwp" event={"ID":"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98","Type":"ContainerStarted","Data":"a61f79e9a6bb7cd95ab84276ad6485d41995ed1040e5ccc126c001a1ecdf78f8"} Apr 22 18:22:08.102707 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:08.102675 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-57cf77bdb6-2vpwp" event={"ID":"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98","Type":"ContainerStarted","Data":"061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff"} Apr 22 18:22:08.121902 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:08.121846 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-57cf77bdb6-2vpwp" podStartSLOduration=1.125294207 podStartE2EDuration="4.121831273s" podCreationTimestamp="2026-04-22 18:22:04 +0000 UTC" firstStartedPulling="2026-04-22 18:22:04.679195227 +0000 UTC m=+81.483258574" lastFinishedPulling="2026-04-22 18:22:07.675732296 +0000 UTC m=+84.479795640" observedRunningTime="2026-04-22 18:22:08.120008054 +0000 UTC m=+84.924071421" watchObservedRunningTime="2026-04-22 18:22:08.121831273 +0000 UTC m=+84.925894639" Apr 22 18:22:14.559346 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:14.559318 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:14.559823 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:14.559408 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:14.564029 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:14.564010 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:15.124574 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:15.124545 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:22:24.028146 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:22:24.028117 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-47xwd" Apr 22 18:23:12.675846 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:12.675805 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:23:12.676448 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:12.676404 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="alertmanager" containerID="cri-o://f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5" gracePeriod=120 Apr 22 18:23:12.676514 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:12.676462 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-metric" containerID="cri-o://5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1" gracePeriod=120 Apr 22 18:23:12.676565 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:12.676516 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-web" containerID="cri-o://1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044" gracePeriod=120 Apr 22 18:23:12.676565 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:12.676504 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="prom-label-proxy" containerID="cri-o://b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7" gracePeriod=120 Apr 22 18:23:12.676661 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:12.676540 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy" containerID="cri-o://a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563" gracePeriod=120 Apr 22 18:23:12.676661 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:12.676540 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="config-reloader" containerID="cri-o://23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98" gracePeriod=120 Apr 22 18:23:13.287248 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287216 2581 generic.go:358] "Generic (PLEG): container finished" podID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerID="b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7" exitCode=0 Apr 22 18:23:13.287248 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287240 2581 generic.go:358] "Generic (PLEG): container finished" podID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerID="a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563" exitCode=0 Apr 22 18:23:13.287248 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287248 2581 generic.go:358] "Generic (PLEG): container finished" podID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerID="23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98" exitCode=0 Apr 22 18:23:13.287446 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287256 2581 generic.go:358] "Generic (PLEG): container finished" podID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerID="f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5" exitCode=0 Apr 22 18:23:13.287446 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287285 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7"} Apr 22 18:23:13.287446 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287316 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563"} Apr 22 18:23:13.287446 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287327 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98"} Apr 22 18:23:13.287446 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.287335 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5"} Apr 22 18:23:13.907435 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.907414 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.944871 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-metric\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.944938 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-main-db\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.944982 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-metrics-client-ca\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945008 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-cluster-tls-config\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945036 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945101 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-web-config\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945125 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945148 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945150 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-web\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.945787 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945672 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-main-db" (OuterVolumeSpecName: "alertmanager-main-db") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "alertmanager-main-db". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 18:23:13.945787 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945718 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 18:23:13.945787 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945778 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-volume\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945833 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9679r\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-kube-api-access-9679r\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945864 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-out\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945892 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.945951 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-tls-assets\") pod \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\" (UID: \"4c77cb4e-4729-4bfa-ac40-e4e0837818f3\") " Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.946037 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle" (OuterVolumeSpecName: "alertmanager-trusted-ca-bundle") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "alertmanager-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.946216 2581 reconciler_common.go:299] "Volume detached for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-trusted-ca-bundle\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.946238 2581 reconciler_common.go:299] "Volume detached for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-alertmanager-main-db\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:13.946474 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.946253 2581 reconciler_common.go:299] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-metrics-client-ca\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:13.952087 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.952057 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:13.952183 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.952135 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-out" (OuterVolumeSpecName: "config-out") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 18:23:13.952309 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.952279 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-kube-api-access-9679r" (OuterVolumeSpecName: "kube-api-access-9679r") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "kube-api-access-9679r". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 18:23:13.952412 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.952392 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 18:23:13.952549 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.952528 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-metric" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-metric") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-metric". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:13.952614 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.952582 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-volume" (OuterVolumeSpecName: "config-volume") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:13.954154 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.954115 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls" (OuterVolumeSpecName: "secret-alertmanager-main-tls") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "secret-alertmanager-main-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:13.954349 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.954325 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-web" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-web") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-web". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:13.955707 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.955681 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-cluster-tls-config" (OuterVolumeSpecName: "cluster-tls-config") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "cluster-tls-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:13.962232 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:13.962209 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-web-config" (OuterVolumeSpecName: "web-config") pod "4c77cb4e-4729-4bfa-ac40-e4e0837818f3" (UID: "4c77cb4e-4729-4bfa-ac40-e4e0837818f3"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:14.046574 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046550 2581 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-metric\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046574 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046573 2581 reconciler_common.go:299] "Volume detached for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-cluster-tls-config\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046603 2581 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046613 2581 reconciler_common.go:299] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-web-config\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046622 2581 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-kube-rbac-proxy-web\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046631 2581 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-volume\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046639 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9679r\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-kube-api-access-9679r\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046648 2581 reconciler_common.go:299] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-config-out\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046656 2581 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-secret-alertmanager-main-tls\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.046691 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.046664 2581 reconciler_common.go:299] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4c77cb4e-4729-4bfa-ac40-e4e0837818f3-tls-assets\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:14.292380 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.292351 2581 generic.go:358] "Generic (PLEG): container finished" podID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerID="5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1" exitCode=0 Apr 22 18:23:14.292380 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.292377 2581 generic.go:358] "Generic (PLEG): container finished" podID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerID="1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044" exitCode=0 Apr 22 18:23:14.292508 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.292395 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1"} Apr 22 18:23:14.292508 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.292420 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044"} Apr 22 18:23:14.292508 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.292430 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4c77cb4e-4729-4bfa-ac40-e4e0837818f3","Type":"ContainerDied","Data":"a39fb4fb56689c82ceed47b3ca569766f21802060dee2320678e96645aa577bd"} Apr 22 18:23:14.292508 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.292444 2581 scope.go:117] "RemoveContainer" containerID="b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7" Apr 22 18:23:14.292508 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.292502 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.300369 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.300336 2581 scope.go:117] "RemoveContainer" containerID="5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1" Apr 22 18:23:14.306885 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.306868 2581 scope.go:117] "RemoveContainer" containerID="a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563" Apr 22 18:23:14.313046 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.313031 2581 scope.go:117] "RemoveContainer" containerID="1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044" Apr 22 18:23:14.317704 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.317678 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:23:14.319855 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.319832 2581 scope.go:117] "RemoveContainer" containerID="23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98" Apr 22 18:23:14.321895 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.321877 2581 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:23:14.326766 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.326748 2581 scope.go:117] "RemoveContainer" containerID="f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5" Apr 22 18:23:14.332767 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.332751 2581 scope.go:117] "RemoveContainer" containerID="1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416" Apr 22 18:23:14.338717 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.338700 2581 scope.go:117] "RemoveContainer" containerID="b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7" Apr 22 18:23:14.338960 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:23:14.338940 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7\": container with ID starting with b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7 not found: ID does not exist" containerID="b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7" Apr 22 18:23:14.339046 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.338971 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7"} err="failed to get container status \"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7\": rpc error: code = NotFound desc = could not find container \"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7\": container with ID starting with b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7 not found: ID does not exist" Apr 22 18:23:14.339046 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339021 2581 scope.go:117] "RemoveContainer" containerID="5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1" Apr 22 18:23:14.339260 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:23:14.339243 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1\": container with ID starting with 5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1 not found: ID does not exist" containerID="5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1" Apr 22 18:23:14.339302 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339266 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1"} err="failed to get container status \"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1\": rpc error: code = NotFound desc = could not find container \"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1\": container with ID starting with 5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1 not found: ID does not exist" Apr 22 18:23:14.339302 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339281 2581 scope.go:117] "RemoveContainer" containerID="a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563" Apr 22 18:23:14.339464 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:23:14.339446 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563\": container with ID starting with a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563 not found: ID does not exist" containerID="a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563" Apr 22 18:23:14.339523 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339473 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563"} err="failed to get container status \"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563\": rpc error: code = NotFound desc = could not find container \"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563\": container with ID starting with a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563 not found: ID does not exist" Apr 22 18:23:14.339523 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339516 2581 scope.go:117] "RemoveContainer" containerID="1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044" Apr 22 18:23:14.339706 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:23:14.339692 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044\": container with ID starting with 1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044 not found: ID does not exist" containerID="1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044" Apr 22 18:23:14.339765 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339712 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044"} err="failed to get container status \"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044\": rpc error: code = NotFound desc = could not find container \"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044\": container with ID starting with 1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044 not found: ID does not exist" Apr 22 18:23:14.339765 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339730 2581 scope.go:117] "RemoveContainer" containerID="23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98" Apr 22 18:23:14.339958 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:23:14.339943 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98\": container with ID starting with 23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98 not found: ID does not exist" containerID="23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98" Apr 22 18:23:14.340022 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339964 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98"} err="failed to get container status \"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98\": rpc error: code = NotFound desc = could not find container \"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98\": container with ID starting with 23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98 not found: ID does not exist" Apr 22 18:23:14.340022 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.339983 2581 scope.go:117] "RemoveContainer" containerID="f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5" Apr 22 18:23:14.340217 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:23:14.340200 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5\": container with ID starting with f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5 not found: ID does not exist" containerID="f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5" Apr 22 18:23:14.340257 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340221 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5"} err="failed to get container status \"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5\": rpc error: code = NotFound desc = could not find container \"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5\": container with ID starting with f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5 not found: ID does not exist" Apr 22 18:23:14.340257 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340235 2581 scope.go:117] "RemoveContainer" containerID="1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416" Apr 22 18:23:14.340441 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:23:14.340423 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416\": container with ID starting with 1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416 not found: ID does not exist" containerID="1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416" Apr 22 18:23:14.340516 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340445 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416"} err="failed to get container status \"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416\": rpc error: code = NotFound desc = could not find container \"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416\": container with ID starting with 1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416 not found: ID does not exist" Apr 22 18:23:14.340516 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340458 2581 scope.go:117] "RemoveContainer" containerID="b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7" Apr 22 18:23:14.340664 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340648 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7"} err="failed to get container status \"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7\": rpc error: code = NotFound desc = could not find container \"b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7\": container with ID starting with b9ba6ef7bf57de6542c5b00891e3c721fc2c52e9155419ee4931da967e4d42b7 not found: ID does not exist" Apr 22 18:23:14.340714 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340665 2581 scope.go:117] "RemoveContainer" containerID="5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1" Apr 22 18:23:14.340866 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340851 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1"} err="failed to get container status \"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1\": rpc error: code = NotFound desc = could not find container \"5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1\": container with ID starting with 5c286ad107ed956e616bc2ceae0b95761a1da63a9d6f51960b74fd12dd9f77e1 not found: ID does not exist" Apr 22 18:23:14.340908 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.340866 2581 scope.go:117] "RemoveContainer" containerID="a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563" Apr 22 18:23:14.341075 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341059 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563"} err="failed to get container status \"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563\": rpc error: code = NotFound desc = could not find container \"a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563\": container with ID starting with a6d9e0c99dc10de7bf9a596415e644b35e680e30240a2d3a4f99c3e0a8481563 not found: ID does not exist" Apr 22 18:23:14.341124 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341076 2581 scope.go:117] "RemoveContainer" containerID="1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044" Apr 22 18:23:14.341279 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341263 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044"} err="failed to get container status \"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044\": rpc error: code = NotFound desc = could not find container \"1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044\": container with ID starting with 1e3fadacb9c70c7e0443562d1c35123fd5cef6b57b6c8e9334ccac814bcc3044 not found: ID does not exist" Apr 22 18:23:14.341279 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341279 2581 scope.go:117] "RemoveContainer" containerID="23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98" Apr 22 18:23:14.341480 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341462 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98"} err="failed to get container status \"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98\": rpc error: code = NotFound desc = could not find container \"23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98\": container with ID starting with 23c2e41b7e9146b3fe1a28905ad084ee510beeff71dfe70fba2ea84cd63eef98 not found: ID does not exist" Apr 22 18:23:14.341534 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341481 2581 scope.go:117] "RemoveContainer" containerID="f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5" Apr 22 18:23:14.341713 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341691 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5"} err="failed to get container status \"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5\": rpc error: code = NotFound desc = could not find container \"f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5\": container with ID starting with f3a616e41aa1dd5db1093d98b8a330b51ee6a47d589050f88e070e3d6ef397d5 not found: ID does not exist" Apr 22 18:23:14.341713 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341713 2581 scope.go:117] "RemoveContainer" containerID="1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416" Apr 22 18:23:14.341970 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.341951 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416"} err="failed to get container status \"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416\": rpc error: code = NotFound desc = could not find container \"1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416\": container with ID starting with 1c330114676ba36a50969990fd8b9ab8e0f0b09565f76468196162e9b5ee5416 not found: ID does not exist" Apr 22 18:23:14.354068 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354048 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:23:14.354311 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354298 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-web" Apr 22 18:23:14.354311 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354311 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-web" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354322 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="alertmanager" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354328 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="alertmanager" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354337 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="config-reloader" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354342 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="config-reloader" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354347 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-metric" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354352 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-metric" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354358 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="init-config-reloader" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354363 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="init-config-reloader" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354371 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="prom-label-proxy" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354375 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="prom-label-proxy" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354385 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354390 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy" Apr 22 18:23:14.354426 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354428 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy" Apr 22 18:23:14.354784 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354436 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="prom-label-proxy" Apr 22 18:23:14.354784 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354443 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="config-reloader" Apr 22 18:23:14.354784 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354449 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="alertmanager" Apr 22 18:23:14.354784 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354455 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-metric" Apr 22 18:23:14.354784 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.354460 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" containerName="kube-rbac-proxy-web" Apr 22 18:23:14.359180 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.359165 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.361875 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.361856 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Apr 22 18:23:14.361988 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.361935 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Apr 22 18:23:14.361988 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.361978 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Apr 22 18:23:14.362092 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.362062 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Apr 22 18:23:14.362092 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.362068 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-2hrm4\"" Apr 22 18:23:14.362242 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.362227 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Apr 22 18:23:14.362316 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.362262 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Apr 22 18:23:14.362316 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.362278 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Apr 22 18:23:14.362414 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.362340 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Apr 22 18:23:14.367763 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.367745 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Apr 22 18:23:14.372590 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.372573 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:23:14.449134 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449114 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/1aef039e-d285-4186-84ff-af200fcccbfd-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449227 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449146 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aef039e-d285-4186-84ff-af200fcccbfd-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449227 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449178 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-config-volume\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449309 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449224 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449309 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449263 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-web-config\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449309 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449296 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449398 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449314 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449398 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449340 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/1aef039e-d285-4186-84ff-af200fcccbfd-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449398 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449365 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1aef039e-d285-4186-84ff-af200fcccbfd-config-out\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449398 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449387 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449553 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449403 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449553 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449427 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5bfl\" (UniqueName: \"kubernetes.io/projected/1aef039e-d285-4186-84ff-af200fcccbfd-kube-api-access-x5bfl\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.449553 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.449457 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1aef039e-d285-4186-84ff-af200fcccbfd-tls-assets\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550405 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550342 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-web-config\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550405 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550372 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550405 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550389 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550583 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550410 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/1aef039e-d285-4186-84ff-af200fcccbfd-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550583 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550426 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1aef039e-d285-4186-84ff-af200fcccbfd-config-out\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550583 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550543 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550728 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550588 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550728 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550624 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x5bfl\" (UniqueName: \"kubernetes.io/projected/1aef039e-d285-4186-84ff-af200fcccbfd-kube-api-access-x5bfl\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550728 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550648 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1aef039e-d285-4186-84ff-af200fcccbfd-tls-assets\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550728 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550715 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/1aef039e-d285-4186-84ff-af200fcccbfd-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550957 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550739 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aef039e-d285-4186-84ff-af200fcccbfd-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550957 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550773 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-config-volume\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.550957 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.550798 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.552459 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.551228 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/1aef039e-d285-4186-84ff-af200fcccbfd-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.552459 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.552276 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/1aef039e-d285-4186-84ff-af200fcccbfd-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.552459 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.552400 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aef039e-d285-4186-84ff-af200fcccbfd-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.553774 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.553583 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-web-config\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.553774 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.553697 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1aef039e-d285-4186-84ff-af200fcccbfd-config-out\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.553950 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.553835 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.554019 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.553963 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.554055 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.554012 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.554116 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.554094 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.555227 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.555208 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.555301 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.555263 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1aef039e-d285-4186-84ff-af200fcccbfd-tls-assets\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.555472 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.555453 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/1aef039e-d285-4186-84ff-af200fcccbfd-config-volume\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.560756 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.560738 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5bfl\" (UniqueName: \"kubernetes.io/projected/1aef039e-d285-4186-84ff-af200fcccbfd-kube-api-access-x5bfl\") pod \"alertmanager-main-0\" (UID: \"1aef039e-d285-4186-84ff-af200fcccbfd\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.667351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.667322 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 18:23:14.793508 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:14.793247 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 18:23:14.795834 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:23:14.795807 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1aef039e_d285_4186_84ff_af200fcccbfd.slice/crio-f8f070ded681c56652f23d0d53267c8d6790d23866679149bed28b448fa1c27c WatchSource:0}: Error finding container f8f070ded681c56652f23d0d53267c8d6790d23866679149bed28b448fa1c27c: Status 404 returned error can't find the container with id f8f070ded681c56652f23d0d53267c8d6790d23866679149bed28b448fa1c27c Apr 22 18:23:15.297301 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:15.297271 2581 generic.go:358] "Generic (PLEG): container finished" podID="1aef039e-d285-4186-84ff-af200fcccbfd" containerID="7536f2f3b343598abef1f134742f7b001d6e5c658d9d8efed6665bf9a35b99de" exitCode=0 Apr 22 18:23:15.297685 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:15.297306 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerDied","Data":"7536f2f3b343598abef1f134742f7b001d6e5c658d9d8efed6665bf9a35b99de"} Apr 22 18:23:15.297685 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:15.297327 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerStarted","Data":"f8f070ded681c56652f23d0d53267c8d6790d23866679149bed28b448fa1c27c"} Apr 22 18:23:15.743172 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:15.743144 2581 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c77cb4e-4729-4bfa-ac40-e4e0837818f3" path="/var/lib/kubelet/pods/4c77cb4e-4729-4bfa-ac40-e4e0837818f3/volumes" Apr 22 18:23:16.302524 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:16.302492 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerStarted","Data":"f60a86c3b71407e3a317ca69f6221f7b1d8b59379beb8166e44a839c80daf71d"} Apr 22 18:23:16.302524 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:16.302523 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerStarted","Data":"ecce54998b8b82487b96822397e9252cfb8be4358bdb36a7df3b648aa080049b"} Apr 22 18:23:16.302901 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:16.302534 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerStarted","Data":"783369bf72887e662cc52fca1ce21dde70c0a53cd592428f8725057a80851043"} Apr 22 18:23:16.302901 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:16.302542 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerStarted","Data":"824d147253400df2bdb1ae6254a6d3d1f62a83ec39385f25b90511e8819196cb"} Apr 22 18:23:16.302901 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:16.302550 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerStarted","Data":"1476bd8d50ed8077bda18475129ed7507495709f908b0f354b2ce07e3f8058d1"} Apr 22 18:23:16.302901 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:16.302557 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"1aef039e-d285-4186-84ff-af200fcccbfd","Type":"ContainerStarted","Data":"00a3876fa467a6f190106b99c2280bfc7b221cdf2c681645fbd2ef41f3294613"} Apr 22 18:23:16.334934 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:16.334869 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.334852341 podStartE2EDuration="2.334852341s" podCreationTimestamp="2026-04-22 18:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 18:23:16.333844331 +0000 UTC m=+153.137907696" watchObservedRunningTime="2026-04-22 18:23:16.334852341 +0000 UTC m=+153.138915699" Apr 22 18:23:23.469351 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.469318 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-54d7977f44-dlx4k"] Apr 22 18:23:23.472282 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.472263 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.491892 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.491863 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54d7977f44-dlx4k"] Apr 22 18:23:23.618256 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.618224 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-oauth-serving-cert\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.618389 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.618262 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-oauth-config\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.618389 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.618293 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-serving-cert\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.618389 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.618349 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-config\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.618389 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.618378 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-trusted-ca-bundle\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.618555 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.618414 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-service-ca\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.618555 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.618499 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf9m6\" (UniqueName: \"kubernetes.io/projected/b4767e67-f922-4d02-8f40-6ca7c16c5286-kube-api-access-mf9m6\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719142 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719111 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mf9m6\" (UniqueName: \"kubernetes.io/projected/b4767e67-f922-4d02-8f40-6ca7c16c5286-kube-api-access-mf9m6\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719249 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719166 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-oauth-serving-cert\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719249 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719192 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-oauth-config\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719249 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719222 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-serving-cert\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719405 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719358 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-config\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719460 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719408 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-trusted-ca-bundle\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719460 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719451 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-service-ca\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.719984 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.719960 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-oauth-serving-cert\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.720073 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.720007 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-config\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.720322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.720301 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-service-ca\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.720357 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.720307 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4767e67-f922-4d02-8f40-6ca7c16c5286-trusted-ca-bundle\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.721747 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.721723 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-oauth-config\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.721896 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.721879 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4767e67-f922-4d02-8f40-6ca7c16c5286-console-serving-cert\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.730428 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.730406 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf9m6\" (UniqueName: \"kubernetes.io/projected/b4767e67-f922-4d02-8f40-6ca7c16c5286-kube-api-access-mf9m6\") pod \"console-54d7977f44-dlx4k\" (UID: \"b4767e67-f922-4d02-8f40-6ca7c16c5286\") " pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.781906 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.781857 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:23.896094 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:23.896056 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54d7977f44-dlx4k"] Apr 22 18:23:23.899247 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:23:23.899219 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4767e67_f922_4d02_8f40_6ca7c16c5286.slice/crio-44bd19aa50d8a7ffd7bd4c6ec49f1b8fec5738391af4b80a09f5f1f80c184eaf WatchSource:0}: Error finding container 44bd19aa50d8a7ffd7bd4c6ec49f1b8fec5738391af4b80a09f5f1f80c184eaf: Status 404 returned error can't find the container with id 44bd19aa50d8a7ffd7bd4c6ec49f1b8fec5738391af4b80a09f5f1f80c184eaf Apr 22 18:23:24.327331 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:24.327295 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54d7977f44-dlx4k" event={"ID":"b4767e67-f922-4d02-8f40-6ca7c16c5286","Type":"ContainerStarted","Data":"fa1de1209b4b3a993977e9450ac4fdb139ca4bf1c7e5e3685c640a94d8fb8f7b"} Apr 22 18:23:24.327331 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:24.327335 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54d7977f44-dlx4k" event={"ID":"b4767e67-f922-4d02-8f40-6ca7c16c5286","Type":"ContainerStarted","Data":"44bd19aa50d8a7ffd7bd4c6ec49f1b8fec5738391af4b80a09f5f1f80c184eaf"} Apr 22 18:23:24.348815 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:24.348764 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54d7977f44-dlx4k" podStartSLOduration=1.348747742 podStartE2EDuration="1.348747742s" podCreationTimestamp="2026-04-22 18:23:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 18:23:24.347804911 +0000 UTC m=+161.151868279" watchObservedRunningTime="2026-04-22 18:23:24.348747742 +0000 UTC m=+161.152811110" Apr 22 18:23:33.782257 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:33.782223 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:33.782257 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:33.782261 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:33.786811 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:33.786789 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:34.359745 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:34.359720 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-54d7977f44-dlx4k" Apr 22 18:23:34.409116 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:34.409088 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-57cf77bdb6-2vpwp"] Apr 22 18:23:59.431945 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.431888 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-57cf77bdb6-2vpwp" podUID="f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" containerName="console" containerID="cri-o://061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff" gracePeriod=15 Apr 22 18:23:59.664876 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.664852 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-57cf77bdb6-2vpwp_f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98/console/0.log" Apr 22 18:23:59.665015 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.664912 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:23:59.784133 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784110 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-service-ca\") pod \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " Apr 22 18:23:59.784278 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784149 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-serving-cert\") pod \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " Apr 22 18:23:59.784278 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784175 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-oauth-config\") pod \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " Apr 22 18:23:59.784278 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784209 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-config\") pod \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " Apr 22 18:23:59.784278 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784235 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-oauth-serving-cert\") pod \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " Apr 22 18:23:59.784278 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784254 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-trusted-ca-bundle\") pod \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " Apr 22 18:23:59.784538 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784296 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnjpq\" (UniqueName: \"kubernetes.io/projected/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-kube-api-access-jnjpq\") pod \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\" (UID: \"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98\") " Apr 22 18:23:59.784589 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784528 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-service-ca" (OuterVolumeSpecName: "service-ca") pod "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" (UID: "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 18:23:59.784652 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784626 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-config" (OuterVolumeSpecName: "console-config") pod "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" (UID: "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 18:23:59.784771 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784746 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" (UID: "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 18:23:59.784771 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.784756 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" (UID: "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 18:23:59.786645 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.786614 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" (UID: "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:59.786749 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.786658 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" (UID: "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 18:23:59.786749 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.786667 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-kube-api-access-jnjpq" (OuterVolumeSpecName: "kube-api-access-jnjpq") pod "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" (UID: "f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98"). InnerVolumeSpecName "kube-api-access-jnjpq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 18:23:59.885735 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.885713 2581 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-oauth-serving-cert\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:59.885735 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.885733 2581 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-trusted-ca-bundle\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:59.885837 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.885742 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-jnjpq\" (UniqueName: \"kubernetes.io/projected/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-kube-api-access-jnjpq\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:59.885837 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.885751 2581 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-service-ca\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:59.885837 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.885761 2581 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-serving-cert\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:59.885837 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.885769 2581 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-oauth-config\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:23:59.885837 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:23:59.885778 2581 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98-console-config\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:24:00.426525 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.426498 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-57cf77bdb6-2vpwp_f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98/console/0.log" Apr 22 18:24:00.426679 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.426542 2581 generic.go:358] "Generic (PLEG): container finished" podID="f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" containerID="061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff" exitCode=2 Apr 22 18:24:00.426679 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.426601 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-57cf77bdb6-2vpwp" Apr 22 18:24:00.426790 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.426602 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-57cf77bdb6-2vpwp" event={"ID":"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98","Type":"ContainerDied","Data":"061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff"} Apr 22 18:24:00.426790 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.426715 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-57cf77bdb6-2vpwp" event={"ID":"f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98","Type":"ContainerDied","Data":"a61f79e9a6bb7cd95ab84276ad6485d41995ed1040e5ccc126c001a1ecdf78f8"} Apr 22 18:24:00.426790 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.426736 2581 scope.go:117] "RemoveContainer" containerID="061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff" Apr 22 18:24:00.435096 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.434874 2581 scope.go:117] "RemoveContainer" containerID="061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff" Apr 22 18:24:00.435348 ip-10-0-141-172 kubenswrapper[2581]: E0422 18:24:00.435112 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff\": container with ID starting with 061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff not found: ID does not exist" containerID="061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff" Apr 22 18:24:00.439246 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.435459 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff"} err="failed to get container status \"061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff\": rpc error: code = NotFound desc = could not find container \"061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff\": container with ID starting with 061ae3eb2deb3c1c02049e3ca88d2019f1b7ce8634df4ef24c9e4568e2ddffff not found: ID does not exist" Apr 22 18:24:00.456304 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.456284 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-57cf77bdb6-2vpwp"] Apr 22 18:24:00.460807 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:00.460788 2581 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-57cf77bdb6-2vpwp"] Apr 22 18:24:01.742740 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:01.742705 2581 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" path="/var/lib/kubelet/pods/f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98/volumes" Apr 22 18:24:07.164124 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.164093 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2ckcp/must-gather-m96g9"] Apr 22 18:24:07.164479 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.164412 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" containerName="console" Apr 22 18:24:07.164479 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.164423 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" containerName="console" Apr 22 18:24:07.164479 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.164477 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="f26a7e6a-9a08-4fe7-98dd-ddbe71e4fb98" containerName="console" Apr 22 18:24:07.168994 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.168979 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.172427 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.172310 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-2ckcp\"/\"kube-root-ca.crt\"" Apr 22 18:24:07.172653 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.172633 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-2ckcp\"/\"openshift-service-ca.crt\"" Apr 22 18:24:07.173039 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.173025 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-2ckcp\"/\"default-dockercfg-47zt7\"" Apr 22 18:24:07.175771 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.175506 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2ckcp/must-gather-m96g9"] Apr 22 18:24:07.241198 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.241173 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a46f7099-8e99-490a-a290-8469fb751f9b-must-gather-output\") pod \"must-gather-m96g9\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.241308 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.241210 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj9bj\" (UniqueName: \"kubernetes.io/projected/a46f7099-8e99-490a-a290-8469fb751f9b-kube-api-access-lj9bj\") pod \"must-gather-m96g9\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.342286 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.342268 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lj9bj\" (UniqueName: \"kubernetes.io/projected/a46f7099-8e99-490a-a290-8469fb751f9b-kube-api-access-lj9bj\") pod \"must-gather-m96g9\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.342372 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.342326 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a46f7099-8e99-490a-a290-8469fb751f9b-must-gather-output\") pod \"must-gather-m96g9\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.342596 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.342583 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a46f7099-8e99-490a-a290-8469fb751f9b-must-gather-output\") pod \"must-gather-m96g9\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.350721 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.350692 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj9bj\" (UniqueName: \"kubernetes.io/projected/a46f7099-8e99-490a-a290-8469fb751f9b-kube-api-access-lj9bj\") pod \"must-gather-m96g9\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.491473 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.491422 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:07.607232 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:07.607098 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2ckcp/must-gather-m96g9"] Apr 22 18:24:07.609892 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:24:07.609868 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda46f7099_8e99_490a_a290_8469fb751f9b.slice/crio-f582d9d61469d69993cb6cc1d0f7dee920fb2ecbadaa0bdfda21d67fbc8dd32d WatchSource:0}: Error finding container f582d9d61469d69993cb6cc1d0f7dee920fb2ecbadaa0bdfda21d67fbc8dd32d: Status 404 returned error can't find the container with id f582d9d61469d69993cb6cc1d0f7dee920fb2ecbadaa0bdfda21d67fbc8dd32d Apr 22 18:24:08.453009 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:08.452955 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2ckcp/must-gather-m96g9" event={"ID":"a46f7099-8e99-490a-a290-8469fb751f9b","Type":"ContainerStarted","Data":"f582d9d61469d69993cb6cc1d0f7dee920fb2ecbadaa0bdfda21d67fbc8dd32d"} Apr 22 18:24:13.469271 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:13.469154 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2ckcp/must-gather-m96g9" event={"ID":"a46f7099-8e99-490a-a290-8469fb751f9b","Type":"ContainerStarted","Data":"f7fde320d69a05a7268ad0a9cf821d8f0fbaf631ee02a96bd6dbcf55059b0ec8"} Apr 22 18:24:13.469271 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:13.469202 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2ckcp/must-gather-m96g9" event={"ID":"a46f7099-8e99-490a-a290-8469fb751f9b","Type":"ContainerStarted","Data":"9d3f51b014906fd57e55d49ff7a3a5480913439a03a3874992700e68ddf75394"} Apr 22 18:24:13.490090 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:13.490045 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2ckcp/must-gather-m96g9" podStartSLOduration=1.067899836 podStartE2EDuration="6.490030419s" podCreationTimestamp="2026-04-22 18:24:07 +0000 UTC" firstStartedPulling="2026-04-22 18:24:07.611521443 +0000 UTC m=+204.415584788" lastFinishedPulling="2026-04-22 18:24:13.033652019 +0000 UTC m=+209.837715371" observedRunningTime="2026-04-22 18:24:13.489100579 +0000 UTC m=+210.293163936" watchObservedRunningTime="2026-04-22 18:24:13.490030419 +0000 UTC m=+210.294093783" Apr 22 18:24:23.497992 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:23.497962 2581 generic.go:358] "Generic (PLEG): container finished" podID="a46f7099-8e99-490a-a290-8469fb751f9b" containerID="9d3f51b014906fd57e55d49ff7a3a5480913439a03a3874992700e68ddf75394" exitCode=0 Apr 22 18:24:23.498400 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:23.498045 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2ckcp/must-gather-m96g9" event={"ID":"a46f7099-8e99-490a-a290-8469fb751f9b","Type":"ContainerDied","Data":"9d3f51b014906fd57e55d49ff7a3a5480913439a03a3874992700e68ddf75394"} Apr 22 18:24:23.498400 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:23.498339 2581 scope.go:117] "RemoveContainer" containerID="9d3f51b014906fd57e55d49ff7a3a5480913439a03a3874992700e68ddf75394" Apr 22 18:24:23.965385 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:23.965309 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2ckcp_must-gather-m96g9_a46f7099-8e99-490a-a290-8469fb751f9b/gather/0.log" Apr 22 18:24:27.253404 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:27.253373 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-x5xw7_f0d1648a-0bbf-4049-b68e-82046b4578f0/global-pull-secret-syncer/0.log" Apr 22 18:24:27.325518 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:27.325487 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-v2tgg_c32b819b-1ae9-498a-876b-87995a803b2c/konnectivity-agent/0.log" Apr 22 18:24:27.371972 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:27.371946 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-141-172.ec2.internal_b128912d2b3d8968717857155603d30b/haproxy/0.log" Apr 22 18:24:29.313135 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.313098 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2ckcp/must-gather-m96g9"] Apr 22 18:24:29.313595 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.313357 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-must-gather-2ckcp/must-gather-m96g9" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" containerName="copy" containerID="cri-o://f7fde320d69a05a7268ad0a9cf821d8f0fbaf631ee02a96bd6dbcf55059b0ec8" gracePeriod=2 Apr 22 18:24:29.314907 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.314881 2581 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2ckcp/must-gather-m96g9"] Apr 22 18:24:29.315872 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.315849 2581 status_manager.go:895] "Failed to get status for pod" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" pod="openshift-must-gather-2ckcp/must-gather-m96g9" err="pods \"must-gather-m96g9\" is forbidden: User \"system:node:ip-10-0-141-172.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-2ckcp\": no relationship found between node 'ip-10-0-141-172.ec2.internal' and this object" Apr 22 18:24:29.516279 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.516254 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2ckcp_must-gather-m96g9_a46f7099-8e99-490a-a290-8469fb751f9b/copy/0.log" Apr 22 18:24:29.516600 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.516569 2581 generic.go:358] "Generic (PLEG): container finished" podID="a46f7099-8e99-490a-a290-8469fb751f9b" containerID="f7fde320d69a05a7268ad0a9cf821d8f0fbaf631ee02a96bd6dbcf55059b0ec8" exitCode=143 Apr 22 18:24:29.535023 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.535000 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2ckcp_must-gather-m96g9_a46f7099-8e99-490a-a290-8469fb751f9b/copy/0.log" Apr 22 18:24:29.535343 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.535330 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:29.537726 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.537706 2581 status_manager.go:895] "Failed to get status for pod" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" pod="openshift-must-gather-2ckcp/must-gather-m96g9" err="pods \"must-gather-m96g9\" is forbidden: User \"system:node:ip-10-0-141-172.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-2ckcp\": no relationship found between node 'ip-10-0-141-172.ec2.internal' and this object" Apr 22 18:24:29.607300 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.607239 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj9bj\" (UniqueName: \"kubernetes.io/projected/a46f7099-8e99-490a-a290-8469fb751f9b-kube-api-access-lj9bj\") pod \"a46f7099-8e99-490a-a290-8469fb751f9b\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " Apr 22 18:24:29.607396 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.607299 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a46f7099-8e99-490a-a290-8469fb751f9b-must-gather-output\") pod \"a46f7099-8e99-490a-a290-8469fb751f9b\" (UID: \"a46f7099-8e99-490a-a290-8469fb751f9b\") " Apr 22 18:24:29.607600 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.607576 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a46f7099-8e99-490a-a290-8469fb751f9b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a46f7099-8e99-490a-a290-8469fb751f9b" (UID: "a46f7099-8e99-490a-a290-8469fb751f9b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 18:24:29.609500 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.609483 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a46f7099-8e99-490a-a290-8469fb751f9b-kube-api-access-lj9bj" (OuterVolumeSpecName: "kube-api-access-lj9bj") pod "a46f7099-8e99-490a-a290-8469fb751f9b" (UID: "a46f7099-8e99-490a-a290-8469fb751f9b"). InnerVolumeSpecName "kube-api-access-lj9bj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 18:24:29.708503 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.708472 2581 reconciler_common.go:299] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a46f7099-8e99-490a-a290-8469fb751f9b-must-gather-output\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:24:29.708503 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.708503 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lj9bj\" (UniqueName: \"kubernetes.io/projected/a46f7099-8e99-490a-a290-8469fb751f9b-kube-api-access-lj9bj\") on node \"ip-10-0-141-172.ec2.internal\" DevicePath \"\"" Apr 22 18:24:29.742934 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:29.742902 2581 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" path="/var/lib/kubelet/pods/a46f7099-8e99-490a-a290-8469fb751f9b/volumes" Apr 22 18:24:30.524320 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.524290 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2ckcp_must-gather-m96g9_a46f7099-8e99-490a-a290-8469fb751f9b/copy/0.log" Apr 22 18:24:30.524735 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.524657 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2ckcp/must-gather-m96g9" Apr 22 18:24:30.524735 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.524650 2581 scope.go:117] "RemoveContainer" containerID="f7fde320d69a05a7268ad0a9cf821d8f0fbaf631ee02a96bd6dbcf55059b0ec8" Apr 22 18:24:30.531484 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.531466 2581 scope.go:117] "RemoveContainer" containerID="9d3f51b014906fd57e55d49ff7a3a5480913439a03a3874992700e68ddf75394" Apr 22 18:24:30.643800 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.643768 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_1aef039e-d285-4186-84ff-af200fcccbfd/alertmanager/0.log" Apr 22 18:24:30.668344 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.668324 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_1aef039e-d285-4186-84ff-af200fcccbfd/config-reloader/0.log" Apr 22 18:24:30.693895 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.693877 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_1aef039e-d285-4186-84ff-af200fcccbfd/kube-rbac-proxy-web/0.log" Apr 22 18:24:30.738359 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.738338 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_1aef039e-d285-4186-84ff-af200fcccbfd/kube-rbac-proxy/0.log" Apr 22 18:24:30.762200 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.762166 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_1aef039e-d285-4186-84ff-af200fcccbfd/kube-rbac-proxy-metric/0.log" Apr 22 18:24:30.787387 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.787335 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_1aef039e-d285-4186-84ff-af200fcccbfd/prom-label-proxy/0.log" Apr 22 18:24:30.813635 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.813617 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_1aef039e-d285-4186-84ff-af200fcccbfd/init-config-reloader/0.log" Apr 22 18:24:30.975200 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:30.975169 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_monitoring-plugin-7dccd58f55-ccnhj_3d87f815-f0b3-4790-bf18-63a31165c49a/monitoring-plugin/0.log" Apr 22 18:24:31.151807 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.151731 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-njfv5_a33be29f-8691-4d49-8d7c-0b6d40636abc/node-exporter/0.log" Apr 22 18:24:31.175941 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.175903 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-njfv5_a33be29f-8691-4d49-8d7c-0b6d40636abc/kube-rbac-proxy/0.log" Apr 22 18:24:31.199311 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.199290 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-njfv5_a33be29f-8691-4d49-8d7c-0b6d40636abc/init-textfile/0.log" Apr 22 18:24:31.609879 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.609850 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-6fc869648c-gt45s_fd1fd636-f128-44d6-9294-d1aad120ef52/telemeter-client/0.log" Apr 22 18:24:31.638433 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.638411 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-6fc869648c-gt45s_fd1fd636-f128-44d6-9294-d1aad120ef52/reload/0.log" Apr 22 18:24:31.664387 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.664357 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-6fc869648c-gt45s_fd1fd636-f128-44d6-9294-d1aad120ef52/kube-rbac-proxy/0.log" Apr 22 18:24:31.698473 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.698453 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-54f58b6d86-znvnl_47c2b707-b606-473b-b666-0b19f7d15ecb/thanos-query/0.log" Apr 22 18:24:31.725355 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.725337 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-54f58b6d86-znvnl_47c2b707-b606-473b-b666-0b19f7d15ecb/kube-rbac-proxy-web/0.log" Apr 22 18:24:31.748865 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.748849 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-54f58b6d86-znvnl_47c2b707-b606-473b-b666-0b19f7d15ecb/kube-rbac-proxy/0.log" Apr 22 18:24:31.773329 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.773310 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-54f58b6d86-znvnl_47c2b707-b606-473b-b666-0b19f7d15ecb/prom-label-proxy/0.log" Apr 22 18:24:31.797865 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.797848 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-54f58b6d86-znvnl_47c2b707-b606-473b-b666-0b19f7d15ecb/kube-rbac-proxy-rules/0.log" Apr 22 18:24:31.823733 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:31.823708 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-54f58b6d86-znvnl_47c2b707-b606-473b-b666-0b19f7d15ecb/kube-rbac-proxy-metrics/0.log" Apr 22 18:24:33.533539 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.533513 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-54d7977f44-dlx4k_b4767e67-f922-4d02-8f40-6ca7c16c5286/console/0.log" Apr 22 18:24:33.647281 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.647244 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg"] Apr 22 18:24:33.647585 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.647569 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" containerName="gather" Apr 22 18:24:33.647662 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.647588 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" containerName="gather" Apr 22 18:24:33.647662 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.647603 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" containerName="copy" Apr 22 18:24:33.647662 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.647611 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" containerName="copy" Apr 22 18:24:33.647865 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.647693 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" containerName="copy" Apr 22 18:24:33.647865 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.647712 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="a46f7099-8e99-490a-a290-8469fb751f9b" containerName="gather" Apr 22 18:24:33.652378 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.652359 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.655111 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.654979 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-tgcct\"/\"kube-root-ca.crt\"" Apr 22 18:24:33.655362 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.655338 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-tgcct\"/\"openshift-service-ca.crt\"" Apr 22 18:24:33.656309 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.656292 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-tgcct\"/\"default-dockercfg-p5vpd\"" Apr 22 18:24:33.658030 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.658009 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg"] Apr 22 18:24:33.738315 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.738276 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sw4g\" (UniqueName: \"kubernetes.io/projected/fc9e73c2-4cdd-498d-beba-c1582a143f28-kube-api-access-9sw4g\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.738315 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.738325 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-podres\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.738528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.738355 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-lib-modules\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.738528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.738404 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-sys\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.738528 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.738447 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-proc\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839067 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.838974 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9sw4g\" (UniqueName: \"kubernetes.io/projected/fc9e73c2-4cdd-498d-beba-c1582a143f28-kube-api-access-9sw4g\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839067 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839049 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-podres\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839132 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-lib-modules\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839200 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-sys\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839261 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-podres\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839322 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839278 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-proc\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839531 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839386 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-proc\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839531 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839402 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-lib-modules\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.839531 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.839440 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc9e73c2-4cdd-498d-beba-c1582a143f28-sys\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.847242 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.847223 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sw4g\" (UniqueName: \"kubernetes.io/projected/fc9e73c2-4cdd-498d-beba-c1582a143f28-kube-api-access-9sw4g\") pod \"perf-node-gather-daemonset-cmdqg\" (UID: \"fc9e73c2-4cdd-498d-beba-c1582a143f28\") " pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:33.963169 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:33.963136 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:34.083037 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.083003 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg"] Apr 22 18:24:34.086474 ip-10-0-141-172 kubenswrapper[2581]: W0422 18:24:34.086450 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podfc9e73c2_4cdd_498d_beba_c1582a143f28.slice/crio-8d6ae8e9311e39f7fe6f721041a70fddaebd0d2bbae73f7359ab2b5c25c98527 WatchSource:0}: Error finding container 8d6ae8e9311e39f7fe6f721041a70fddaebd0d2bbae73f7359ab2b5c25c98527: Status 404 returned error can't find the container with id 8d6ae8e9311e39f7fe6f721041a70fddaebd0d2bbae73f7359ab2b5c25c98527 Apr 22 18:24:34.537189 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.537154 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" event={"ID":"fc9e73c2-4cdd-498d-beba-c1582a143f28","Type":"ContainerStarted","Data":"600dac2a931e3938c448a961c554479e8ec3d54e64f5df99f95849e3239725b9"} Apr 22 18:24:34.537189 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.537190 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" event={"ID":"fc9e73c2-4cdd-498d-beba-c1582a143f28","Type":"ContainerStarted","Data":"8d6ae8e9311e39f7fe6f721041a70fddaebd0d2bbae73f7359ab2b5c25c98527"} Apr 22 18:24:34.537544 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.537283 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:34.553379 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.553338 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" podStartSLOduration=1.553327711 podStartE2EDuration="1.553327711s" podCreationTimestamp="2026-04-22 18:24:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 18:24:34.55166122 +0000 UTC m=+231.355724585" watchObservedRunningTime="2026-04-22 18:24:34.553327711 +0000 UTC m=+231.357391076" Apr 22 18:24:34.618512 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.618486 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-57ft4_e39bb5e3-7f87-49f6-8be9-68b1fefc9b02/dns/0.log" Apr 22 18:24:34.645936 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.645890 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-57ft4_e39bb5e3-7f87-49f6-8be9-68b1fefc9b02/kube-rbac-proxy/0.log" Apr 22 18:24:34.770700 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:34.770670 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-pqljk_f0bc6045-fff5-4c2b-a848-e95a54d2aaf8/dns-node-resolver/0.log" Apr 22 18:24:35.192606 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:35.192572 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_image-registry-f6646996-g48x4_0326059a-7ea7-44ae-bc43-7619240760e7/registry/0.log" Apr 22 18:24:35.221171 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:35.221140 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-qxft5_bbb98e98-69b1-42c6-a9d1-250ac4d831f3/node-ca/0.log" Apr 22 18:24:36.425384 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:36.425353 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-b9bkd_629aaa70-005e-4af9-9ce5-ca15d667f5e7/serve-healthcheck-canary/0.log" Apr 22 18:24:36.930258 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:36.930233 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-sh9jw_8315b63f-0468-47cb-b528-7534d99126a4/kube-rbac-proxy/0.log" Apr 22 18:24:36.975628 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:36.975606 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-sh9jw_8315b63f-0468-47cb-b528-7534d99126a4/exporter/0.log" Apr 22 18:24:37.017122 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:37.017102 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-sh9jw_8315b63f-0468-47cb-b528-7534d99126a4/extractor/0.log" Apr 22 18:24:40.550124 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:40.550096 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-tgcct/perf-node-gather-daemonset-cmdqg" Apr 22 18:24:43.023891 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.023824 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-x2fkn_c0602c54-63d4-4509-bf26-1bfa53e55b60/kube-multus-additional-cni-plugins/0.log" Apr 22 18:24:43.049723 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.049696 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-x2fkn_c0602c54-63d4-4509-bf26-1bfa53e55b60/egress-router-binary-copy/0.log" Apr 22 18:24:43.074871 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.074842 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-x2fkn_c0602c54-63d4-4509-bf26-1bfa53e55b60/cni-plugins/0.log" Apr 22 18:24:43.101467 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.101441 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-x2fkn_c0602c54-63d4-4509-bf26-1bfa53e55b60/bond-cni-plugin/0.log" Apr 22 18:24:43.129871 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.129843 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-x2fkn_c0602c54-63d4-4509-bf26-1bfa53e55b60/routeoverride-cni/0.log" Apr 22 18:24:43.160150 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.160121 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-x2fkn_c0602c54-63d4-4509-bf26-1bfa53e55b60/whereabouts-cni-bincopy/0.log" Apr 22 18:24:43.188197 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.188173 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-x2fkn_c0602c54-63d4-4509-bf26-1bfa53e55b60/whereabouts-cni/0.log" Apr 22 18:24:43.286680 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.286606 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p27q9_e6d25e3f-fd86-49df-bcb4-b6ac18c164e1/kube-multus/0.log" Apr 22 18:24:43.418898 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.418876 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-xl5pv_b5ac120e-7d32-45b8-a5d6-57aa386e732f/network-metrics-daemon/0.log" Apr 22 18:24:43.448875 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:43.448847 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-xl5pv_b5ac120e-7d32-45b8-a5d6-57aa386e732f/kube-rbac-proxy/0.log" Apr 22 18:24:44.188861 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.188831 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-controller/0.log" Apr 22 18:24:44.211735 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.211714 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-acl-logging/0.log" Apr 22 18:24:44.213026 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.213003 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovn-acl-logging/1.log" Apr 22 18:24:44.236777 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.236748 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/kube-rbac-proxy-node/0.log" Apr 22 18:24:44.274857 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.274831 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/kube-rbac-proxy-ovn-metrics/0.log" Apr 22 18:24:44.298758 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.298734 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/northd/0.log" Apr 22 18:24:44.323195 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.323178 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/nbdb/0.log" Apr 22 18:24:44.349541 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.349493 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/sbdb/0.log" Apr 22 18:24:44.442808 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:44.442779 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-44fqr_90936ce6-5c5f-4f38-9487-0e4631ff77e6/ovnkube-controller/0.log" Apr 22 18:24:46.122871 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:46.122839 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-47xwd_e4946d3e-bc1b-4f87-8fb8-db35971cf837/network-check-target-container/0.log" Apr 22 18:24:47.126665 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:47.126638 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-96txh_a0ad3985-82e5-4fed-89ff-cd6f30087d98/iptables-alerter/0.log" Apr 22 18:24:48.007473 ip-10-0-141-172 kubenswrapper[2581]: I0422 18:24:48.007447 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-zhjmm_721979eb-ddad-4268-a95a-833cf12d6074/tuned/0.log"