Apr 17 09:18:23.962821 ip-10-0-137-152 systemd[1]: kubelet.service: Failed to load environment files: No such file or directory Apr 17 09:18:23.962831 ip-10-0-137-152 systemd[1]: kubelet.service: Failed to run 'start-pre' task: No such file or directory Apr 17 09:18:23.962840 ip-10-0-137-152 systemd[1]: kubelet.service: Failed with result 'resources'. Apr 17 09:18:23.963131 ip-10-0-137-152 systemd[1]: Failed to start Kubernetes Kubelet. Apr 17 09:18:34.212177 ip-10-0-137-152 systemd[1]: kubelet.service: Failed to schedule restart job: Unit crio.service not found. Apr 17 09:18:34.212192 ip-10-0-137-152 systemd[1]: kubelet.service: Failed with result 'resources'. -- Boot bbca9ecf4c9c41cf9482a1a4c9ae524f -- Apr 17 09:21:07.051114 ip-10-0-137-152 systemd[1]: Starting Kubernetes Kubelet... Apr 17 09:21:07.456733 ip-10-0-137-152 kubenswrapper[2569]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 09:21:07.456733 ip-10-0-137-152 kubenswrapper[2569]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 17 09:21:07.456733 ip-10-0-137-152 kubenswrapper[2569]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 09:21:07.456733 ip-10-0-137-152 kubenswrapper[2569]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 17 09:21:07.456733 ip-10-0-137-152 kubenswrapper[2569]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 09:21:07.457813 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.457729 2569 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 17 09:21:07.462585 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462570 2569 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:21:07.462585 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462585 2569 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462589 2569 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462593 2569 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462595 2569 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462598 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462601 2569 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462604 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462606 2569 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462609 2569 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462611 2569 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462614 2569 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462617 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462620 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462623 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462625 2569 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462628 2569 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462630 2569 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462634 2569 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462638 2569 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462641 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:21:07.462652 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462649 2569 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462652 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462655 2569 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462657 2569 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462660 2569 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462663 2569 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462666 2569 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462668 2569 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462671 2569 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462674 2569 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462676 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462679 2569 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462681 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462684 2569 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462687 2569 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462689 2569 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462692 2569 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462694 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462697 2569 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462699 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:21:07.463151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462702 2569 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462704 2569 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462707 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462709 2569 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462711 2569 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462714 2569 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462716 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462719 2569 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462721 2569 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462724 2569 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462726 2569 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462729 2569 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462731 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462735 2569 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462738 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462741 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462743 2569 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462746 2569 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462749 2569 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462752 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:21:07.463686 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462755 2569 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462758 2569 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462760 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462763 2569 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462765 2569 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462768 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462771 2569 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462773 2569 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462776 2569 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462780 2569 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462784 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462787 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462789 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462792 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462795 2569 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462797 2569 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462800 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462802 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462805 2569 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462809 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:21:07.464188 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462811 2569 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462828 2569 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462831 2569 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462834 2569 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.462836 2569 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463242 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463248 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463251 2569 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463253 2569 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463256 2569 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463259 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463261 2569 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463264 2569 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463267 2569 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463270 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463272 2569 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463275 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463278 2569 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463280 2569 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:21:07.464668 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463283 2569 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463286 2569 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463289 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463291 2569 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463294 2569 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463296 2569 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463299 2569 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463301 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463304 2569 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463306 2569 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463310 2569 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463312 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463317 2569 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463321 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463324 2569 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463327 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463330 2569 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463332 2569 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463335 2569 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:21:07.465141 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463338 2569 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463342 2569 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463347 2569 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463349 2569 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463352 2569 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463354 2569 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463357 2569 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463359 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463362 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463365 2569 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463368 2569 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463370 2569 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463373 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463375 2569 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463378 2569 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463381 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463383 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463386 2569 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463388 2569 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463391 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:21:07.465602 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463393 2569 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463395 2569 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463398 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463400 2569 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463403 2569 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463405 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463408 2569 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463410 2569 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463413 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463415 2569 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463418 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463421 2569 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463423 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463426 2569 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463429 2569 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463432 2569 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463434 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463436 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463439 2569 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463442 2569 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:21:07.466133 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463445 2569 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463448 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463451 2569 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463453 2569 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463455 2569 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463458 2569 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463460 2569 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463463 2569 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463465 2569 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463468 2569 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463471 2569 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463473 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.463475 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464078 2569 flags.go:64] FLAG: --address="0.0.0.0" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464088 2569 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464094 2569 flags.go:64] FLAG: --anonymous-auth="true" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464099 2569 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464103 2569 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464106 2569 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464111 2569 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464115 2569 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 17 09:21:07.466638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464119 2569 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464122 2569 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464126 2569 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464129 2569 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464132 2569 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464136 2569 flags.go:64] FLAG: --cgroup-root="" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464139 2569 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464142 2569 flags.go:64] FLAG: --client-ca-file="" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464145 2569 flags.go:64] FLAG: --cloud-config="" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464148 2569 flags.go:64] FLAG: --cloud-provider="external" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464151 2569 flags.go:64] FLAG: --cluster-dns="[]" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464155 2569 flags.go:64] FLAG: --cluster-domain="" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464159 2569 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464162 2569 flags.go:64] FLAG: --config-dir="" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464165 2569 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464168 2569 flags.go:64] FLAG: --container-log-max-files="5" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464172 2569 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464175 2569 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464177 2569 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464181 2569 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464184 2569 flags.go:64] FLAG: --contention-profiling="false" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464186 2569 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464189 2569 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464193 2569 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464196 2569 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 17 09:21:07.467160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464200 2569 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464203 2569 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464206 2569 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464209 2569 flags.go:64] FLAG: --enable-load-reader="false" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464212 2569 flags.go:64] FLAG: --enable-server="true" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464215 2569 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464220 2569 flags.go:64] FLAG: --event-burst="100" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464223 2569 flags.go:64] FLAG: --event-qps="50" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464226 2569 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464229 2569 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464232 2569 flags.go:64] FLAG: --eviction-hard="" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464236 2569 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464239 2569 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464242 2569 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464245 2569 flags.go:64] FLAG: --eviction-soft="" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464248 2569 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464251 2569 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464254 2569 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464259 2569 flags.go:64] FLAG: --experimental-mounter-path="" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464262 2569 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464265 2569 flags.go:64] FLAG: --fail-swap-on="true" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464268 2569 flags.go:64] FLAG: --feature-gates="" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464272 2569 flags.go:64] FLAG: --file-check-frequency="20s" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464275 2569 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464279 2569 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 17 09:21:07.467784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464282 2569 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464285 2569 flags.go:64] FLAG: --healthz-port="10248" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464288 2569 flags.go:64] FLAG: --help="false" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464291 2569 flags.go:64] FLAG: --hostname-override="ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464294 2569 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464298 2569 flags.go:64] FLAG: --http-check-frequency="20s" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464301 2569 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464305 2569 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464308 2569 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464311 2569 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464313 2569 flags.go:64] FLAG: --image-service-endpoint="" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464316 2569 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464319 2569 flags.go:64] FLAG: --kube-api-burst="100" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464322 2569 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464325 2569 flags.go:64] FLAG: --kube-api-qps="50" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464328 2569 flags.go:64] FLAG: --kube-reserved="" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464331 2569 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464334 2569 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464338 2569 flags.go:64] FLAG: --kubelet-cgroups="" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464341 2569 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464344 2569 flags.go:64] FLAG: --lock-file="" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464347 2569 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464350 2569 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464353 2569 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 17 09:21:07.468406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464359 2569 flags.go:64] FLAG: --log-json-split-stream="false" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464363 2569 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464366 2569 flags.go:64] FLAG: --log-text-split-stream="false" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464369 2569 flags.go:64] FLAG: --logging-format="text" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464372 2569 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464376 2569 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464379 2569 flags.go:64] FLAG: --manifest-url="" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464382 2569 flags.go:64] FLAG: --manifest-url-header="" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464386 2569 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464389 2569 flags.go:64] FLAG: --max-open-files="1000000" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464393 2569 flags.go:64] FLAG: --max-pods="110" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464396 2569 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464399 2569 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464402 2569 flags.go:64] FLAG: --memory-manager-policy="None" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464405 2569 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464408 2569 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464411 2569 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464414 2569 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464421 2569 flags.go:64] FLAG: --node-status-max-images="50" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464424 2569 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464427 2569 flags.go:64] FLAG: --oom-score-adj="-999" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464430 2569 flags.go:64] FLAG: --pod-cidr="" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464433 2569 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 17 09:21:07.468994 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464439 2569 flags.go:64] FLAG: --pod-manifest-path="" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464442 2569 flags.go:64] FLAG: --pod-max-pids="-1" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464445 2569 flags.go:64] FLAG: --pods-per-core="0" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464448 2569 flags.go:64] FLAG: --port="10250" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464451 2569 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464454 2569 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-00e46a67cad85d224" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464457 2569 flags.go:64] FLAG: --qos-reserved="" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464460 2569 flags.go:64] FLAG: --read-only-port="10255" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464463 2569 flags.go:64] FLAG: --register-node="true" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464466 2569 flags.go:64] FLAG: --register-schedulable="true" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464470 2569 flags.go:64] FLAG: --register-with-taints="" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464474 2569 flags.go:64] FLAG: --registry-burst="10" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464477 2569 flags.go:64] FLAG: --registry-qps="5" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464480 2569 flags.go:64] FLAG: --reserved-cpus="" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464482 2569 flags.go:64] FLAG: --reserved-memory="" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464486 2569 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464489 2569 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464492 2569 flags.go:64] FLAG: --rotate-certificates="false" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464495 2569 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464498 2569 flags.go:64] FLAG: --runonce="false" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464500 2569 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464503 2569 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464506 2569 flags.go:64] FLAG: --seccomp-default="false" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464509 2569 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464512 2569 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464515 2569 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 17 09:21:07.469543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464518 2569 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464521 2569 flags.go:64] FLAG: --storage-driver-password="root" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464524 2569 flags.go:64] FLAG: --storage-driver-secure="false" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464527 2569 flags.go:64] FLAG: --storage-driver-table="stats" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464530 2569 flags.go:64] FLAG: --storage-driver-user="root" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464532 2569 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464535 2569 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464538 2569 flags.go:64] FLAG: --system-cgroups="" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464543 2569 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464548 2569 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464551 2569 flags.go:64] FLAG: --tls-cert-file="" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464554 2569 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464557 2569 flags.go:64] FLAG: --tls-min-version="" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464561 2569 flags.go:64] FLAG: --tls-private-key-file="" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464564 2569 flags.go:64] FLAG: --topology-manager-policy="none" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464567 2569 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464572 2569 flags.go:64] FLAG: --topology-manager-scope="container" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464575 2569 flags.go:64] FLAG: --v="2" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464579 2569 flags.go:64] FLAG: --version="false" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464583 2569 flags.go:64] FLAG: --vmodule="" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464587 2569 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464590 2569 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464676 2569 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464680 2569 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464683 2569 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:21:07.470191 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464686 2569 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464689 2569 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464692 2569 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464695 2569 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464697 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464700 2569 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464702 2569 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464705 2569 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464708 2569 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464711 2569 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464714 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464716 2569 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464719 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464722 2569 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464724 2569 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464727 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464730 2569 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464732 2569 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464735 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:21:07.470794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464737 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464740 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464742 2569 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464745 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464748 2569 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464751 2569 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464753 2569 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464756 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464758 2569 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464761 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464763 2569 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464767 2569 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464770 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464773 2569 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464776 2569 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464778 2569 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464782 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464784 2569 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464788 2569 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:21:07.471575 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464791 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464795 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464798 2569 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464800 2569 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464803 2569 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464805 2569 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464808 2569 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464810 2569 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464824 2569 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464827 2569 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464830 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464832 2569 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464835 2569 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464838 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464840 2569 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464843 2569 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464846 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464849 2569 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464852 2569 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464854 2569 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:21:07.472151 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464857 2569 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464859 2569 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464862 2569 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464864 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464867 2569 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464869 2569 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464872 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464874 2569 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464877 2569 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464879 2569 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464882 2569 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464884 2569 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464887 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464889 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464892 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464895 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464897 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464900 2569 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464902 2569 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464905 2569 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:21:07.472639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464907 2569 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:21:07.473142 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464910 2569 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:21:07.473142 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464913 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:21:07.473142 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464915 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:21:07.473142 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.464918 2569 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:21:07.473142 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.464926 2569 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 09:21:07.473286 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.473267 2569 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 17 09:21:07.473319 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.473287 2569 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 17 09:21:07.473347 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473338 2569 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:21:07.473347 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473343 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:21:07.473347 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473347 2569 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473351 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473354 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473357 2569 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473360 2569 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473363 2569 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473366 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473369 2569 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473371 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473374 2569 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473377 2569 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473380 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473382 2569 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473385 2569 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473387 2569 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473390 2569 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473392 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473395 2569 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473398 2569 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473400 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:21:07.473452 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473404 2569 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473407 2569 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473410 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473413 2569 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473417 2569 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473421 2569 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473425 2569 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473428 2569 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473430 2569 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473433 2569 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473436 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473439 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473442 2569 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473444 2569 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473448 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473451 2569 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473453 2569 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473456 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473459 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:21:07.473963 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473462 2569 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473465 2569 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473467 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473470 2569 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473473 2569 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473475 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473479 2569 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473481 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473484 2569 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473487 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473490 2569 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473492 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473495 2569 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473498 2569 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473500 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473503 2569 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473505 2569 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473508 2569 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473511 2569 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473513 2569 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:21:07.474429 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473516 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473519 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473521 2569 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473524 2569 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473527 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473530 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473533 2569 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473536 2569 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473538 2569 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473541 2569 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473544 2569 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473547 2569 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473551 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473554 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473556 2569 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473560 2569 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473562 2569 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473565 2569 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473567 2569 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473570 2569 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:21:07.474947 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473572 2569 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473575 2569 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473578 2569 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473580 2569 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473583 2569 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.473588 2569 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473685 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473690 2569 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473693 2569 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473696 2569 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473699 2569 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473702 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473705 2569 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473707 2569 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473710 2569 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:21:07.475465 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473713 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473715 2569 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473718 2569 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473721 2569 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473724 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473727 2569 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473729 2569 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473732 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473735 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473738 2569 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473740 2569 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473743 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473746 2569 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473749 2569 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473751 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473754 2569 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473757 2569 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473759 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473762 2569 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473764 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:21:07.475844 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473767 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473769 2569 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473772 2569 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473774 2569 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473777 2569 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473779 2569 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473782 2569 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473784 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473787 2569 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473789 2569 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473792 2569 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473794 2569 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473797 2569 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473800 2569 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473802 2569 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473805 2569 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473808 2569 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473811 2569 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473830 2569 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:21:07.476327 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473835 2569 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473839 2569 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473842 2569 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473846 2569 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473849 2569 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473852 2569 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473855 2569 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473858 2569 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473861 2569 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473863 2569 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473866 2569 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473868 2569 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473872 2569 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473875 2569 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473878 2569 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473880 2569 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473883 2569 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473885 2569 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473888 2569 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473890 2569 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:21:07.476797 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473893 2569 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473895 2569 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473898 2569 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473901 2569 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473904 2569 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473906 2569 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473909 2569 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473911 2569 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473914 2569 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473917 2569 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473919 2569 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473922 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473924 2569 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473927 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473930 2569 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473932 2569 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473935 2569 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:21:07.477300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:07.473937 2569 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:21:07.477713 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.473942 2569 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 09:21:07.477713 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.474702 2569 server.go:962] "Client rotation is on, will bootstrap in background" Apr 17 09:21:07.478259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.478244 2569 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 17 09:21:07.479118 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.479107 2569 server.go:1019] "Starting client certificate rotation" Apr 17 09:21:07.479225 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.479208 2569 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 17 09:21:07.479260 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.479247 2569 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 17 09:21:07.505044 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.505026 2569 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 17 09:21:07.509220 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.509201 2569 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 17 09:21:07.520091 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.520071 2569 log.go:25] "Validated CRI v1 runtime API" Apr 17 09:21:07.525095 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.525081 2569 log.go:25] "Validated CRI v1 image API" Apr 17 09:21:07.527596 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.527580 2569 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 17 09:21:07.531507 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.531486 2569 fs.go:135] Filesystem UUIDs: map[0b8c306f-2b4e-424d-83f5-153214fb2f99:/dev/nvme0n1p3 2627fa69-7ad4-4249-8f40-947416238d0c:/dev/nvme0n1p4 7B77-95E7:/dev/nvme0n1p2] Apr 17 09:21:07.531573 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.531506 2569 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 17 09:21:07.534841 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.534809 2569 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 17 09:21:07.538369 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.538260 2569 manager.go:217] Machine: {Timestamp:2026-04-17 09:21:07.536353271 +0000 UTC m=+0.377390474 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3100243 MemoryCapacity:33164492800 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2eba19f19fc73f6537b4b5a1d965f5 SystemUUID:ec2eba19-f19f-c73f-6537-b4b5a1d965f5 BootID:bbca9ecf-4c9c-41cf-9482-a1a4c9ae524f Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16582246400 Type:vfs Inodes:4048400 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6632898560 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16582246400 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:8a:d0:d1:2b:1b Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:8a:d0:d1:2b:1b Speed:0 Mtu:9001} {Name:ovs-system MacAddress:5a:2d:4a:0a:8a:3e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33164492800 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 17 09:21:07.538369 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.538361 2569 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 17 09:21:07.538495 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.538438 2569 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 17 09:21:07.540126 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.540084 2569 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 17 09:21:07.540520 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.540127 2569 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-137-152.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 17 09:21:07.540575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.540531 2569 topology_manager.go:138] "Creating topology manager with none policy" Apr 17 09:21:07.540575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.540540 2569 container_manager_linux.go:306] "Creating device plugin manager" Apr 17 09:21:07.540575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.540553 2569 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 17 09:21:07.541278 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.541266 2569 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 17 09:21:07.542617 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.542607 2569 state_mem.go:36] "Initialized new in-memory state store" Apr 17 09:21:07.542738 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.542729 2569 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 17 09:21:07.545243 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.545234 2569 kubelet.go:491] "Attempting to sync node with API server" Apr 17 09:21:07.546035 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.546026 2569 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 17 09:21:07.546075 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.546045 2569 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 17 09:21:07.546075 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.546054 2569 kubelet.go:397] "Adding apiserver pod source" Apr 17 09:21:07.546075 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.546062 2569 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 17 09:21:07.547155 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.547132 2569 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 17 09:21:07.547214 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.547162 2569 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 17 09:21:07.549874 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.549859 2569 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 17 09:21:07.551098 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.551085 2569 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 17 09:21:07.551166 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.551151 2569 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-4l26c" Apr 17 09:21:07.552971 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.552958 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.552975 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.552982 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.552989 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.552995 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.553001 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.553009 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.553015 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 17 09:21:07.553024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.553023 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 17 09:21:07.553236 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.553029 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 17 09:21:07.553236 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.553038 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 17 09:21:07.553236 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.553046 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 17 09:21:07.554811 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.554801 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 17 09:21:07.554811 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.554811 2569 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 17 09:21:07.557712 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.557693 2569 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-137-152.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 17 09:21:07.557810 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.557760 2569 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-137-152.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 17 09:21:07.557894 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.557853 2569 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 17 09:21:07.558354 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.558342 2569 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 17 09:21:07.558395 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.558375 2569 server.go:1295] "Started kubelet" Apr 17 09:21:07.558528 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.558485 2569 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 17 09:21:07.558576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.558546 2569 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 17 09:21:07.558663 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.558580 2569 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 17 09:21:07.559087 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.559064 2569 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-4l26c" Apr 17 09:21:07.559388 ip-10-0-137-152 systemd[1]: Started Kubernetes Kubelet. Apr 17 09:21:07.559588 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.559574 2569 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 17 09:21:07.560395 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.560351 2569 server.go:317] "Adding debug handlers to kubelet server" Apr 17 09:21:07.565091 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.565071 2569 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 17 09:21:07.565624 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.565607 2569 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 17 09:21:07.566484 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.566447 2569 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 17 09:21:07.566484 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.566466 2569 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 17 09:21:07.566673 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.566657 2569 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 17 09:21:07.566753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.566737 2569 reconstruct.go:97] "Volume reconstruction finished" Apr 17 09:21:07.566753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.566749 2569 reconciler.go:26] "Reconciler: start to sync state" Apr 17 09:21:07.567046 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.567009 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:07.567658 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.567644 2569 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 17 09:21:07.567658 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.567657 2569 factory.go:55] Registering systemd factory Apr 17 09:21:07.567797 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.567664 2569 factory.go:223] Registration of the systemd container factory successfully Apr 17 09:21:07.568494 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.568469 2569 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:21:07.568836 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.568805 2569 factory.go:153] Registering CRI-O factory Apr 17 09:21:07.568836 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.568834 2569 factory.go:223] Registration of the crio container factory successfully Apr 17 09:21:07.568993 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.568856 2569 factory.go:103] Registering Raw factory Apr 17 09:21:07.568993 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.568870 2569 manager.go:1196] Started watching for new ooms in manager Apr 17 09:21:07.569718 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.569697 2569 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 17 09:21:07.570097 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.570082 2569 manager.go:319] Starting recovery of all containers Apr 17 09:21:07.571124 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.571076 2569 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-10-0-137-152.ec2.internal\" not found" node="ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.581879 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.581684 2569 manager.go:324] Recovery completed Apr 17 09:21:07.585741 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.585729 2569 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:21:07.588531 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.588517 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:21:07.588593 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.588543 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:21:07.588593 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.588556 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:21:07.589000 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.588987 2569 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 17 09:21:07.589000 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.588999 2569 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 17 09:21:07.589087 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.589015 2569 state_mem.go:36] "Initialized new in-memory state store" Apr 17 09:21:07.592146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.592127 2569 policy_none.go:49] "None policy: Start" Apr 17 09:21:07.592146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.592143 2569 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 17 09:21:07.592218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.592153 2569 state_mem.go:35] "Initializing new in-memory state store" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.632899 2569 manager.go:341] "Starting Device Plugin manager" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.632929 2569 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.632939 2569 server.go:85] "Starting device plugin registration server" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.633168 2569 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.633182 2569 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.633269 2569 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.633354 2569 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.633363 2569 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.633860 2569 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 17 09:21:07.654936 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.633898 2569 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:07.700634 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.700611 2569 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 17 09:21:07.701839 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.701810 2569 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 17 09:21:07.701889 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.701851 2569 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 17 09:21:07.701889 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.701868 2569 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 17 09:21:07.701889 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.701876 2569 kubelet.go:2451] "Starting kubelet main sync loop" Apr 17 09:21:07.702012 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.701907 2569 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 17 09:21:07.704311 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.704295 2569 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:21:07.734143 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.734105 2569 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:21:07.735154 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.735139 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:21:07.735211 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.735169 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:21:07.735211 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.735180 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:21:07.735211 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.735202 2569 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.744757 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.744743 2569 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.744796 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.744764 2569 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-137-152.ec2.internal\": node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:07.760862 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.760844 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:07.802165 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.802144 2569 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal"] Apr 17 09:21:07.802218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.802204 2569 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:21:07.802920 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.802905 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:21:07.802996 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.802938 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:21:07.802996 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.802950 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:21:07.804998 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.804987 2569 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:21:07.805142 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805129 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.805177 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805157 2569 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:21:07.805688 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805667 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:21:07.805763 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805697 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:21:07.805763 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805671 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:21:07.805763 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805731 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:21:07.805763 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805751 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:21:07.805763 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.805710 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:21:07.807769 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.807754 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.807855 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.807780 2569 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:21:07.808392 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.808377 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:21:07.808470 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.808406 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:21:07.808470 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.808429 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:21:07.833359 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.833340 2569 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-137-152.ec2.internal\" not found" node="ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.837645 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.837630 2569 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-137-152.ec2.internal\" not found" node="ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.861220 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.861197 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:07.868500 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.868482 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/6669a9b7ffc9643b9e141a323972a5e6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal\" (UID: \"6669a9b7ffc9643b9e141a323972a5e6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.868560 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.868505 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6669a9b7ffc9643b9e141a323972a5e6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal\" (UID: \"6669a9b7ffc9643b9e141a323972a5e6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.868560 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.868526 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/311677e33e183475ac3960f0885188ab-config\") pod \"kube-apiserver-proxy-ip-10-0-137-152.ec2.internal\" (UID: \"311677e33e183475ac3960f0885188ab\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.961989 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:07.961956 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:07.969339 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.969317 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/6669a9b7ffc9643b9e141a323972a5e6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal\" (UID: \"6669a9b7ffc9643b9e141a323972a5e6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.969390 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.969284 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/6669a9b7ffc9643b9e141a323972a5e6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal\" (UID: \"6669a9b7ffc9643b9e141a323972a5e6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.969390 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.969385 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6669a9b7ffc9643b9e141a323972a5e6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal\" (UID: \"6669a9b7ffc9643b9e141a323972a5e6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.969453 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.969402 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/311677e33e183475ac3960f0885188ab-config\") pod \"kube-apiserver-proxy-ip-10-0-137-152.ec2.internal\" (UID: \"311677e33e183475ac3960f0885188ab\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.969453 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.969425 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/311677e33e183475ac3960f0885188ab-config\") pod \"kube-apiserver-proxy-ip-10-0-137-152.ec2.internal\" (UID: \"311677e33e183475ac3960f0885188ab\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" Apr 17 09:21:07.969505 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:07.969455 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6669a9b7ffc9643b9e141a323972a5e6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal\" (UID: \"6669a9b7ffc9643b9e141a323972a5e6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:08.062748 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.062682 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.135176 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.135153 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:08.140689 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.140667 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" Apr 17 09:21:08.163797 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.163770 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.264283 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.264237 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.364770 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.364698 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.406765 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.406741 2569 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:21:08.464844 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.464806 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.479178 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.479155 2569 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 17 09:21:08.479277 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.479262 2569 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 17 09:21:08.479319 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.479299 2569 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 17 09:21:08.479319 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.479308 2569 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 17 09:21:08.561388 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.561354 2569 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-16 09:16:07 +0000 UTC" deadline="2027-09-14 10:00:35.884050639 +0000 UTC" Apr 17 09:21:08.561388 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.561384 2569 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="12360h39m27.322669861s" Apr 17 09:21:08.565124 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.565101 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.565219 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.565171 2569 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 17 09:21:08.574999 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.574982 2569 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 17 09:21:08.600934 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.600914 2569 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-9t89j" Apr 17 09:21:08.610641 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.610624 2569 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-9t89j" Apr 17 09:21:08.666190 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.666132 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.710623 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:08.710594 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6669a9b7ffc9643b9e141a323972a5e6.slice/crio-e0fec3ce4bee62b7f79725d2487bfff3b2ef68da8f47cb3080defcaa7d580552 WatchSource:0}: Error finding container e0fec3ce4bee62b7f79725d2487bfff3b2ef68da8f47cb3080defcaa7d580552: Status 404 returned error can't find the container with id e0fec3ce4bee62b7f79725d2487bfff3b2ef68da8f47cb3080defcaa7d580552 Apr 17 09:21:08.711048 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:08.711029 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod311677e33e183475ac3960f0885188ab.slice/crio-77756dc6c8a8b4823affb34cffc77d3c9b6eaabd67d2cb3e4335ba532dd56efc WatchSource:0}: Error finding container 77756dc6c8a8b4823affb34cffc77d3c9b6eaabd67d2cb3e4335ba532dd56efc: Status 404 returned error can't find the container with id 77756dc6c8a8b4823affb34cffc77d3c9b6eaabd67d2cb3e4335ba532dd56efc Apr 17 09:21:08.714996 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.714980 2569 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 09:21:08.766595 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.766573 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.866990 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.866959 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.967394 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:08.967326 2569 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-152.ec2.internal\" not found" Apr 17 09:21:08.979160 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:08.979134 2569 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:21:09.067112 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.067089 2569 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" Apr 17 09:21:09.077894 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.077872 2569 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 17 09:21:09.078765 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.078746 2569 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" Apr 17 09:21:09.084614 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.084598 2569 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 17 09:21:09.333649 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.333561 2569 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:21:09.434417 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.434388 2569 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:21:09.547037 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.547011 2569 apiserver.go:52] "Watching apiserver" Apr 17 09:21:09.553782 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.553759 2569 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 17 09:21:09.555580 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.555558 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-9bwlv","openshift-multus/multus-additional-cni-plugins-g7bkg","openshift-multus/network-metrics-daemon-4kbg5","openshift-network-operator/iptables-alerter-dd7tv","openshift-cluster-node-tuning-operator/tuned-rjm7r","openshift-dns/node-resolver-b5dkm","openshift-image-registry/node-ca-t5ckx","openshift-network-diagnostics/network-check-target-2cw78","openshift-ovn-kubernetes/ovnkube-node-gd2rv","kube-system/konnectivity-agent-b65w4","kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal"] Apr 17 09:21:09.564275 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.564256 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.566655 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.566634 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-ldmbl\"" Apr 17 09:21:09.566655 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.566648 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 17 09:21:09.566801 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.566651 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 17 09:21:09.566801 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.566641 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 17 09:21:09.568563 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.568546 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.568664 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.568604 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:09.568726 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.568669 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:09.571005 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.570978 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-4xddh\"" Apr 17 09:21:09.571110 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.571055 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 17 09:21:09.571168 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.571154 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 17 09:21:09.571367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.571344 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 17 09:21:09.571458 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.571371 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 17 09:21:09.571458 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.571353 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 17 09:21:09.573324 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.573247 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.575174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.575142 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-nfd8z\"" Apr 17 09:21:09.575305 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.575225 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:21:09.575305 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.575253 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 17 09:21:09.575444 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.575317 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 17 09:21:09.575648 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.575625 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.575975 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.575743 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.576773 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.576752 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.576871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.576785 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09912815-97d7-47ef-939f-9f3815c992c1-host\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.576871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.576808 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.576980 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.576901 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsbcr\" (UniqueName: \"kubernetes.io/projected/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-kube-api-access-fsbcr\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.576980 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.576938 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx6gk\" (UniqueName: \"kubernetes.io/projected/833e0b85-0aba-4f60-bbf6-a215ed02831a-kube-api-access-jx6gk\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:09.576980 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.576971 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09912815-97d7-47ef-939f-9f3815c992c1-serviceca\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.577139 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.576998 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89c2c\" (UniqueName: \"kubernetes.io/projected/09912815-97d7-47ef-939f-9f3815c992c1-kube-api-access-89c2c\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.577139 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577022 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-system-cni-dir\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.577139 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577060 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-os-release\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.577139 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577088 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cni-binary-copy\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.577139 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577110 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-tuning-conf-dir\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.577360 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577152 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:09.577360 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577184 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cnibin\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.577459 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577387 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-gw65g\"" Apr 17 09:21:09.577737 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577721 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-5zcxm\"" Apr 17 09:21:09.577855 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577744 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 17 09:21:09.577855 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.577775 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 17 09:21:09.578106 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.578073 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:21:09.578474 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.578452 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 17 09:21:09.578544 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.578487 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.580260 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.580217 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 17 09:21:09.581406 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.580938 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:09.581406 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.581107 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:09.581574 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.581448 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-fc98p\"" Apr 17 09:21:09.584520 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.584470 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.586503 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.586480 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 17 09:21:09.586608 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.586490 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-s2p4j\"" Apr 17 09:21:09.586880 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.586861 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.587347 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.587326 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 17 09:21:09.587519 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.587503 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 17 09:21:09.587519 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.587516 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 17 09:21:09.587757 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.587740 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 17 09:21:09.587875 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.587832 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 17 09:21:09.588689 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.588675 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 17 09:21:09.588784 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.588727 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 17 09:21:09.588857 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.588840 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-tvg9s\"" Apr 17 09:21:09.589120 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.589103 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.590949 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.590932 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 17 09:21:09.591147 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.591129 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 17 09:21:09.591268 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.591247 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-m5qnc\"" Apr 17 09:21:09.591362 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.591311 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 17 09:21:09.611374 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.611342 2569 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-16 09:16:08 +0000 UTC" deadline="2027-10-03 00:10:12.872132785 +0000 UTC" Apr 17 09:21:09.611374 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.611372 2569 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="12806h49m3.260764031s" Apr 17 09:21:09.667892 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.667870 2569 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.677917 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-os-release\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.677955 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-tuning-conf-dir\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.677986 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovnkube-config\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678012 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/904ead54-94c1-4e84-9851-0c8506ed7bbc-kube-api-access-hr9p4\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678042 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cnibin\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678051 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-os-release\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678068 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9fcd330e-f463-420a-a91c-f2a6e23daebc-tmp-dir\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678092 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678116 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-cni-bin\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678123 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-tuning-conf-dir\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678132 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cnibin\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678140 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-tmp\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.678188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678177 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/d7b21dc1-e456-4f3a-827a-6d17899801ad-agent-certs\") pod \"konnectivity-agent-b65w4\" (UID: \"d7b21dc1-e456-4f3a-827a-6d17899801ad\") " pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678204 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-cni-netd\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678231 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-os-release\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678255 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-cni-bin\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678279 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-kubelet\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678300 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysctl-conf\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678339 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-systemd\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678361 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-lib-modules\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678387 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-conf-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678415 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-multus-certs\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678463 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/d7b21dc1-e456-4f3a-827a-6d17899801ad-konnectivity-ca\") pod \"konnectivity-agent-b65w4\" (UID: \"d7b21dc1-e456-4f3a-827a-6d17899801ad\") " pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678513 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-var-lib-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678542 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-etc-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678575 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09912815-97d7-47ef-939f-9f3815c992c1-serviceca\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678607 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-system-cni-dir\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678642 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678693 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-system-cni-dir\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.678833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678764 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-systemd-units\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.678834 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678863 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-sys\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.678961 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:21:10.178915368 +0000 UTC m=+3.019952581 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.678998 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dct2\" (UniqueName: \"kubernetes.io/projected/e686455d-e53d-4a81-b3c4-74eb05f3da00-kube-api-access-5dct2\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679054 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-slash\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679078 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-node-log\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679099 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwhvw\" (UniqueName: \"kubernetes.io/projected/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-kube-api-access-dwhvw\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679109 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09912815-97d7-47ef-939f-9f3815c992c1-serviceca\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679141 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-socket-dir-parent\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679171 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-hostroot\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679184 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-etc-kubernetes\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679208 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-kubelet-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679255 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-etc-selinux\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679301 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679374 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-netns\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.679575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679405 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-socket-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679431 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovnkube-script-lib\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679463 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-var-lib-kubelet\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679492 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-89c2c\" (UniqueName: \"kubernetes.io/projected/09912815-97d7-47ef-939f-9f3815c992c1-kube-api-access-89c2c\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679538 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cni-binary-copy\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679564 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnhp4\" (UniqueName: \"kubernetes.io/projected/9fcd330e-f463-420a-a91c-f2a6e23daebc-kube-api-access-hnhp4\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679589 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-ovn\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679607 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-run-ovn-kubernetes\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679631 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-env-overrides\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679659 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-cni-multus\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679736 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-run\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679788 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e686455d-e53d-4a81-b3c4-74eb05f3da00-cni-binary-copy\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679812 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-run-netns\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679876 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysctl-d\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.679953 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-kubelet\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680006 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-registration-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.680344 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680035 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f0ca441-19bb-4396-adb5-c2b223622c08-host-slash\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680065 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680101 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jx6gk\" (UniqueName: \"kubernetes.io/projected/833e0b85-0aba-4f60-bbf6-a215ed02831a-kube-api-access-jx6gk\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680133 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-systemd\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680151 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cni-binary-copy\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680173 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680207 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-system-cni-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680233 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-cnibin\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680265 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-sys-fs\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680302 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-cni-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680337 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9fcd330e-f463-420a-a91c-f2a6e23daebc-hosts-file\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680358 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-log-socket\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680381 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-modprobe-d\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680403 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-device-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680428 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfv5v\" (UniqueName: \"kubernetes.io/projected/2f0ca441-19bb-4396-adb5-c2b223622c08-kube-api-access-dfv5v\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680459 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovn-node-metrics-cert\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.681086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680483 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-kubernetes\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680513 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680539 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09912815-97d7-47ef-939f-9f3815c992c1-host\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680566 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysconfig\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680589 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-tuned\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680622 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggsn6\" (UniqueName: \"kubernetes.io/projected/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-kube-api-access-ggsn6\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680646 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-daemon-config\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680674 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fsbcr\" (UniqueName: \"kubernetes.io/projected/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-kube-api-access-fsbcr\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680651 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09912815-97d7-47ef-939f-9f3815c992c1-host\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680708 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-host\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680758 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-k8s-cni-cncf-io\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680785 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/2f0ca441-19bb-4396-adb5-c2b223622c08-iptables-alerter-script\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.680991 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.681633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.681236 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.692670 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.692644 2569 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 17 09:21:09.696345 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.696323 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-89c2c\" (UniqueName: \"kubernetes.io/projected/09912815-97d7-47ef-939f-9f3815c992c1-kube-api-access-89c2c\") pod \"node-ca-t5ckx\" (UID: \"09912815-97d7-47ef-939f-9f3815c992c1\") " pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.696420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.696323 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsbcr\" (UniqueName: \"kubernetes.io/projected/70fa4740-d7b3-4bc0-8829-d52ad2ac1144-kube-api-access-fsbcr\") pod \"multus-additional-cni-plugins-g7bkg\" (UID: \"70fa4740-d7b3-4bc0-8829-d52ad2ac1144\") " pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.696420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.696326 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx6gk\" (UniqueName: \"kubernetes.io/projected/833e0b85-0aba-4f60-bbf6-a215ed02831a-kube-api-access-jx6gk\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:09.708558 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.708516 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" event={"ID":"6669a9b7ffc9643b9e141a323972a5e6","Type":"ContainerStarted","Data":"e0fec3ce4bee62b7f79725d2487bfff3b2ef68da8f47cb3080defcaa7d580552"} Apr 17 09:21:09.709733 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.709710 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" event={"ID":"311677e33e183475ac3960f0885188ab","Type":"ContainerStarted","Data":"77756dc6c8a8b4823affb34cffc77d3c9b6eaabd67d2cb3e4335ba532dd56efc"} Apr 17 09:21:09.781419 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781387 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-ovn\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.781419 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781422 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-run-ovn-kubernetes\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781440 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-env-overrides\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781457 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-cni-multus\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781484 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-run\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781505 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-ovn\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781518 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-run-ovn-kubernetes\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781544 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-run\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781566 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-cni-multus\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781571 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e686455d-e53d-4a81-b3c4-74eb05f3da00-cni-binary-copy\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.781636 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781612 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-run-netns\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781652 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysctl-d\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781657 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-run-netns\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781690 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-kubelet\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781743 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-registration-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781767 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f0ca441-19bb-4396-adb5-c2b223622c08-host-slash\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781791 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysctl-d\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781841 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-registration-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781844 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-systemd\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781848 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-kubelet\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781878 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-systemd\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781880 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781901 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781904 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f0ca441-19bb-4396-adb5-c2b223622c08-host-slash\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781916 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-system-cni-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781942 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-cnibin\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781960 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-system-cni-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782053 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781966 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-sys-fs\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781982 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-env-overrides\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781991 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-cni-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.781987 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-cnibin\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782014 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9fcd330e-f463-420a-a91c-f2a6e23daebc-hosts-file\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782022 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-sys-fs\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782036 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-log-socket\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782058 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-cni-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782062 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-modprobe-d\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782065 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-log-socket\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782091 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-device-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782113 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dfv5v\" (UniqueName: \"kubernetes.io/projected/2f0ca441-19bb-4396-adb5-c2b223622c08-kube-api-access-dfv5v\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782119 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9fcd330e-f463-420a-a91c-f2a6e23daebc-hosts-file\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782151 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovn-node-metrics-cert\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782157 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-modprobe-d\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782162 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-device-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782171 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-kubernetes\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782194 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysconfig\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.782700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782214 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-tuned\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782243 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ggsn6\" (UniqueName: \"kubernetes.io/projected/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-kube-api-access-ggsn6\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782271 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-daemon-config\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782279 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysconfig\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782357 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-kubernetes\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782295 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-host\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782406 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-k8s-cni-cncf-io\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782416 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-host\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782429 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/2f0ca441-19bb-4396-adb5-c2b223622c08-iptables-alerter-script\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782455 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovnkube-config\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782456 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-k8s-cni-cncf-io\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782480 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/904ead54-94c1-4e84-9851-0c8506ed7bbc-kube-api-access-hr9p4\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782507 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9fcd330e-f463-420a-a91c-f2a6e23daebc-tmp-dir\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782533 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782593 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-cni-bin\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782622 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-tmp\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782645 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/d7b21dc1-e456-4f3a-827a-6d17899801ad-agent-certs\") pod \"konnectivity-agent-b65w4\" (UID: \"d7b21dc1-e456-4f3a-827a-6d17899801ad\") " pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782672 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-cni-netd\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.783576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782729 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-cni-bin\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782696 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-os-release\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782773 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-cni-bin\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782789 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-os-release\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782802 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-kubelet\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782845 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysctl-conf\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782870 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-systemd\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782882 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-run-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782896 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-lib-modules\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782920 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-conf-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782944 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-multus-certs\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782956 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/2f0ca441-19bb-4396-adb5-c2b223622c08-iptables-alerter-script\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782977 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/d7b21dc1-e456-4f3a-827a-6d17899801ad-konnectivity-ca\") pod \"konnectivity-agent-b65w4\" (UID: \"d7b21dc1-e456-4f3a-827a-6d17899801ad\") " pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.782958 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovnkube-config\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783001 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-var-lib-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783038 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-systemd\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783043 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-etc-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783075 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-var-lib-cni-bin\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.784606 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783088 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-systemd-units\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783117 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-sys\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783120 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-kubelet\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783145 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5dct2\" (UniqueName: \"kubernetes.io/projected/e686455d-e53d-4a81-b3c4-74eb05f3da00-kube-api-access-5dct2\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783170 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-slash\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783192 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-node-log\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783203 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9fcd330e-f463-420a-a91c-f2a6e23daebc-tmp-dir\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783222 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dwhvw\" (UniqueName: \"kubernetes.io/projected/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-kube-api-access-dwhvw\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783236 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-sysctl-conf\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783254 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-socket-dir-parent\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783260 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-etc-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783277 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-multus-certs\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783282 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-hostroot\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783307 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-etc-kubernetes\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783353 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-kubelet-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783378 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-etc-selinux\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783404 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783409 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-lib-modules\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.785379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783440 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-netns\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783470 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-socket-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783492 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-conf-dir\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783496 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovnkube-script-lib\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783621 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-var-lib-kubelet\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783640 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-systemd-units\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783658 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hnhp4\" (UniqueName: \"kubernetes.io/projected/9fcd330e-f463-420a-a91c-f2a6e23daebc-kube-api-access-hnhp4\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783697 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-sys\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783849 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/d7b21dc1-e456-4f3a-827a-6d17899801ad-konnectivity-ca\") pod \"konnectivity-agent-b65w4\" (UID: \"d7b21dc1-e456-4f3a-827a-6d17899801ad\") " pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783914 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-var-lib-kubelet\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783919 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-host-run-netns\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783958 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-var-lib-openvswitch\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783970 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovnkube-script-lib\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783974 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-slash\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.783972 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-host-cni-netd\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784001 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-node-log\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784026 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-etc-kubernetes\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784047 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-etc-selinux\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.786199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784075 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-socket-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784076 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-hostroot\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784116 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/904ead54-94c1-4e84-9851-0c8506ed7bbc-kubelet-dir\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784197 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-socket-dir-parent\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784606 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e686455d-e53d-4a81-b3c4-74eb05f3da00-cni-binary-copy\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784771 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-etc-tuned\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.784925 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e686455d-e53d-4a81-b3c4-74eb05f3da00-multus-daemon-config\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.785244 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-ovn-node-metrics-cert\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.785581 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/d7b21dc1-e456-4f3a-827a-6d17899801ad-agent-certs\") pod \"konnectivity-agent-b65w4\" (UID: \"d7b21dc1-e456-4f3a-827a-6d17899801ad\") " pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.787094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.786203 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-tmp\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.788213 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.788193 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:09.788313 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.788224 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:09.788313 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.788237 2569 projected.go:194] Error preparing data for projected volume kube-api-access-rh6ms for pod openshift-network-diagnostics/network-check-target-2cw78: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:09.788313 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:09.788300 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms podName:e338cd11-dfc1-43bb-8092-32e10f1e62a5 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:10.288281366 +0000 UTC m=+3.129318557 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-rh6ms" (UniqueName: "kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms") pod "network-check-target-2cw78" (UID: "e338cd11-dfc1-43bb-8092-32e10f1e62a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:09.790526 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.790506 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggsn6\" (UniqueName: \"kubernetes.io/projected/0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32-kube-api-access-ggsn6\") pod \"tuned-rjm7r\" (UID: \"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32\") " pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.790642 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.790621 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/904ead54-94c1-4e84-9851-0c8506ed7bbc-kube-api-access-hr9p4\") pod \"aws-ebs-csi-driver-node-8l4wz\" (UID: \"904ead54-94c1-4e84-9851-0c8506ed7bbc\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:09.791358 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.791313 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfv5v\" (UniqueName: \"kubernetes.io/projected/2f0ca441-19bb-4396-adb5-c2b223622c08-kube-api-access-dfv5v\") pod \"iptables-alerter-dd7tv\" (UID: \"2f0ca441-19bb-4396-adb5-c2b223622c08\") " pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.791871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.791798 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dct2\" (UniqueName: \"kubernetes.io/projected/e686455d-e53d-4a81-b3c4-74eb05f3da00-kube-api-access-5dct2\") pod \"multus-9bwlv\" (UID: \"e686455d-e53d-4a81-b3c4-74eb05f3da00\") " pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.791871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.791810 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwhvw\" (UniqueName: \"kubernetes.io/projected/1ea35608-b7a5-4952-8b6c-f24a85b0ff89-kube-api-access-dwhvw\") pod \"ovnkube-node-gd2rv\" (UID: \"1ea35608-b7a5-4952-8b6c-f24a85b0ff89\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.791871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.791793 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnhp4\" (UniqueName: \"kubernetes.io/projected/9fcd330e-f463-420a-a91c-f2a6e23daebc-kube-api-access-hnhp4\") pod \"node-resolver-b5dkm\" (UID: \"9fcd330e-f463-420a-a91c-f2a6e23daebc\") " pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.875038 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.874963 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-t5ckx" Apr 17 09:21:09.883616 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.883594 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" Apr 17 09:21:09.893150 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.893133 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-dd7tv" Apr 17 09:21:09.897703 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.897673 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" Apr 17 09:21:09.905227 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.905210 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-b5dkm" Apr 17 09:21:09.911704 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.911689 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9bwlv" Apr 17 09:21:09.917271 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.917254 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:09.923846 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.923830 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:09.928302 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:09.928281 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" Apr 17 09:21:10.186444 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.186383 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:10.186557 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:10.186492 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:10.186557 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:10.186541 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:21:11.18652558 +0000 UTC m=+4.027562770 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:10.236700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.236674 2569 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:21:10.371439 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.371415 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode686455d_e53d_4a81_b3c4_74eb05f3da00.slice/crio-e50520b08656209c6f14549597d0afdf9a6b4476499dc38cbfbd5c44d3c736b9 WatchSource:0}: Error finding container e50520b08656209c6f14549597d0afdf9a6b4476499dc38cbfbd5c44d3c736b9: Status 404 returned error can't find the container with id e50520b08656209c6f14549597d0afdf9a6b4476499dc38cbfbd5c44d3c736b9 Apr 17 09:21:10.372425 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.372395 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fcd330e_f463_420a_a91c_f2a6e23daebc.slice/crio-c3af6584f26943296f73a949bee59119fa0cfb8ab2ff44e1a2e8a10fecee1fea WatchSource:0}: Error finding container c3af6584f26943296f73a949bee59119fa0cfb8ab2ff44e1a2e8a10fecee1fea: Status 404 returned error can't find the container with id c3af6584f26943296f73a949bee59119fa0cfb8ab2ff44e1a2e8a10fecee1fea Apr 17 09:21:10.373955 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.373933 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod904ead54_94c1_4e84_9851_0c8506ed7bbc.slice/crio-745dd097e2d666650d58885ada0400ad0722f373f85c77613da2034031b2d9ed WatchSource:0}: Error finding container 745dd097e2d666650d58885ada0400ad0722f373f85c77613da2034031b2d9ed: Status 404 returned error can't find the container with id 745dd097e2d666650d58885ada0400ad0722f373f85c77613da2034031b2d9ed Apr 17 09:21:10.378330 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.378293 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f0ca441_19bb_4396_adb5_c2b223622c08.slice/crio-a143bda0b0e545a662546d9117bdf6635f7d5b43e60afbc396a318a05f069f3a WatchSource:0}: Error finding container a143bda0b0e545a662546d9117bdf6635f7d5b43e60afbc396a318a05f069f3a: Status 404 returned error can't find the container with id a143bda0b0e545a662546d9117bdf6635f7d5b43e60afbc396a318a05f069f3a Apr 17 09:21:10.378473 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.378457 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09912815_97d7_47ef_939f_9f3815c992c1.slice/crio-0a63ab8ab078bc2f9bdb6c331481bb0ef4b6501cd7dac097f32ec2b31d7a5619 WatchSource:0}: Error finding container 0a63ab8ab078bc2f9bdb6c331481bb0ef4b6501cd7dac097f32ec2b31d7a5619: Status 404 returned error can't find the container with id 0a63ab8ab078bc2f9bdb6c331481bb0ef4b6501cd7dac097f32ec2b31d7a5619 Apr 17 09:21:10.387431 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.387403 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:10.387560 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:10.387539 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:10.387674 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:10.387564 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:10.387674 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:10.387578 2569 projected.go:194] Error preparing data for projected volume kube-api-access-rh6ms for pod openshift-network-diagnostics/network-check-target-2cw78: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:10.387674 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:10.387652 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms podName:e338cd11-dfc1-43bb-8092-32e10f1e62a5 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:11.387626311 +0000 UTC m=+4.228663521 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-rh6ms" (UniqueName: "kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms") pod "network-check-target-2cw78" (UID: "e338cd11-dfc1-43bb-8092-32e10f1e62a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:10.399718 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.399680 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70fa4740_d7b3_4bc0_8829_d52ad2ac1144.slice/crio-4822706a1a7b1ec9e87f6191234fb552b4bc32579c5e6f65b018b18dca939937 WatchSource:0}: Error finding container 4822706a1a7b1ec9e87f6191234fb552b4bc32579c5e6f65b018b18dca939937: Status 404 returned error can't find the container with id 4822706a1a7b1ec9e87f6191234fb552b4bc32579c5e6f65b018b18dca939937 Apr 17 09:21:10.400378 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.400358 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ea35608_b7a5_4952_8b6c_f24a85b0ff89.slice/crio-588071018ce13efc1d8bc28e3a5af6c002c18f3609179c9d3f17be547708cd04 WatchSource:0}: Error finding container 588071018ce13efc1d8bc28e3a5af6c002c18f3609179c9d3f17be547708cd04: Status 404 returned error can't find the container with id 588071018ce13efc1d8bc28e3a5af6c002c18f3609179c9d3f17be547708cd04 Apr 17 09:21:10.401112 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.401093 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7b21dc1_e456_4f3a_827a_6d17899801ad.slice/crio-b67afa14e9ba37fb9447ae503170bd828fcf8f1eae2cadebed46f9d47c463ccd WatchSource:0}: Error finding container b67afa14e9ba37fb9447ae503170bd828fcf8f1eae2cadebed46f9d47c463ccd: Status 404 returned error can't find the container with id b67afa14e9ba37fb9447ae503170bd828fcf8f1eae2cadebed46f9d47c463ccd Apr 17 09:21:10.401919 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:21:10.401898 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ed27ada_6c47_4d7b_9fe8_b0a6afea7e32.slice/crio-222b1874a2701e92c8349f7d785084a9341162f12083bcd25d2cb11c72c5527f WatchSource:0}: Error finding container 222b1874a2701e92c8349f7d785084a9341162f12083bcd25d2cb11c72c5527f: Status 404 returned error can't find the container with id 222b1874a2701e92c8349f7d785084a9341162f12083bcd25d2cb11c72c5527f Apr 17 09:21:10.612616 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.612391 2569 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-16 09:16:08 +0000 UTC" deadline="2027-11-08 15:16:48.033653127 +0000 UTC" Apr 17 09:21:10.612616 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.612616 2569 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="13685h55m37.42104298s" Apr 17 09:21:10.714676 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.714595 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerStarted","Data":"4822706a1a7b1ec9e87f6191234fb552b4bc32579c5e6f65b018b18dca939937"} Apr 17 09:21:10.716918 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.716889 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" event={"ID":"904ead54-94c1-4e84-9851-0c8506ed7bbc","Type":"ContainerStarted","Data":"745dd097e2d666650d58885ada0400ad0722f373f85c77613da2034031b2d9ed"} Apr 17 09:21:10.717879 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.717853 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-dd7tv" event={"ID":"2f0ca441-19bb-4396-adb5-c2b223622c08","Type":"ContainerStarted","Data":"a143bda0b0e545a662546d9117bdf6635f7d5b43e60afbc396a318a05f069f3a"} Apr 17 09:21:10.718785 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.718759 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-b5dkm" event={"ID":"9fcd330e-f463-420a-a91c-f2a6e23daebc","Type":"ContainerStarted","Data":"c3af6584f26943296f73a949bee59119fa0cfb8ab2ff44e1a2e8a10fecee1fea"} Apr 17 09:21:10.719866 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.719843 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9bwlv" event={"ID":"e686455d-e53d-4a81-b3c4-74eb05f3da00","Type":"ContainerStarted","Data":"e50520b08656209c6f14549597d0afdf9a6b4476499dc38cbfbd5c44d3c736b9"} Apr 17 09:21:10.721576 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.721553 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" event={"ID":"311677e33e183475ac3960f0885188ab","Type":"ContainerStarted","Data":"ea0438426744f257775150110ebb9eb11c4260219e9fb0783b412f6447993a96"} Apr 17 09:21:10.723175 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.723153 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-t5ckx" event={"ID":"09912815-97d7-47ef-939f-9f3815c992c1","Type":"ContainerStarted","Data":"0a63ab8ab078bc2f9bdb6c331481bb0ef4b6501cd7dac097f32ec2b31d7a5619"} Apr 17 09:21:10.724324 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.724301 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" event={"ID":"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32","Type":"ContainerStarted","Data":"222b1874a2701e92c8349f7d785084a9341162f12083bcd25d2cb11c72c5527f"} Apr 17 09:21:10.725464 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.725377 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-b65w4" event={"ID":"d7b21dc1-e456-4f3a-827a-6d17899801ad","Type":"ContainerStarted","Data":"b67afa14e9ba37fb9447ae503170bd828fcf8f1eae2cadebed46f9d47c463ccd"} Apr 17 09:21:10.727441 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:10.727347 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"588071018ce13efc1d8bc28e3a5af6c002c18f3609179c9d3f17be547708cd04"} Apr 17 09:21:11.193485 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:11.192857 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:11.193485 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.193042 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:11.193485 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.193125 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:21:13.193082256 +0000 UTC m=+6.034119472 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:11.395603 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:11.394906 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:11.395603 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.395118 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:11.395603 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.395142 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:11.395603 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.395154 2569 projected.go:194] Error preparing data for projected volume kube-api-access-rh6ms for pod openshift-network-diagnostics/network-check-target-2cw78: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:11.395603 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.395217 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms podName:e338cd11-dfc1-43bb-8092-32e10f1e62a5 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:13.395197152 +0000 UTC m=+6.236234345 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-rh6ms" (UniqueName: "kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms") pod "network-check-target-2cw78" (UID: "e338cd11-dfc1-43bb-8092-32e10f1e62a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:11.702917 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:11.702170 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:11.702917 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.702312 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:11.702917 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:11.702410 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:11.702917 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:11.702478 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:11.738690 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:11.738657 2569 generic.go:358] "Generic (PLEG): container finished" podID="6669a9b7ffc9643b9e141a323972a5e6" containerID="cb2d259825f58276ecb2617f37db2901d931336e1652bde161dbc3781e0e4688" exitCode=0 Apr 17 09:21:11.738902 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:11.738877 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" event={"ID":"6669a9b7ffc9643b9e141a323972a5e6","Type":"ContainerDied","Data":"cb2d259825f58276ecb2617f37db2901d931336e1652bde161dbc3781e0e4688"} Apr 17 09:21:11.753066 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:11.753021 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-152.ec2.internal" podStartSLOduration=2.753004533 podStartE2EDuration="2.753004533s" podCreationTimestamp="2026-04-17 09:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:21:10.735657599 +0000 UTC m=+3.576694812" watchObservedRunningTime="2026-04-17 09:21:11.753004533 +0000 UTC m=+4.594041745" Apr 17 09:21:12.753067 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:12.752381 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" event={"ID":"6669a9b7ffc9643b9e141a323972a5e6","Type":"ContainerStarted","Data":"cc0e4c3f080b217cf13e2fa88dc1c117dad25c91b061f4217c7d7f574a4078ee"} Apr 17 09:21:12.771550 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:12.771493 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-152.ec2.internal" podStartSLOduration=3.771473426 podStartE2EDuration="3.771473426s" podCreationTimestamp="2026-04-17 09:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:21:12.771215226 +0000 UTC m=+5.612252440" watchObservedRunningTime="2026-04-17 09:21:12.771473426 +0000 UTC m=+5.612510644" Apr 17 09:21:13.211064 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:13.210970 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:13.211228 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.211165 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:13.211286 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.211229 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:21:17.211210014 +0000 UTC m=+10.052247223 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:13.412698 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:13.412661 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:13.412886 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.412837 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:13.412886 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.412866 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:13.412886 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.412881 2569 projected.go:194] Error preparing data for projected volume kube-api-access-rh6ms for pod openshift-network-diagnostics/network-check-target-2cw78: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:13.413072 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.412938 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms podName:e338cd11-dfc1-43bb-8092-32e10f1e62a5 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:17.412922925 +0000 UTC m=+10.253960120 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-rh6ms" (UniqueName: "kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms") pod "network-check-target-2cw78" (UID: "e338cd11-dfc1-43bb-8092-32e10f1e62a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:13.706240 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:13.706163 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:13.706393 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.706290 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:13.706561 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:13.706540 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:13.706691 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:13.706670 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:15.702835 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:15.702780 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:15.703312 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:15.702941 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:15.703596 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:15.703573 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:15.703730 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:15.703703 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:17.244631 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:17.244555 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:17.245202 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.244720 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:17.245202 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.244794 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:21:25.244772524 +0000 UTC m=+18.085809716 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:17.445920 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:17.445882 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:17.446098 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.446046 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:17.446098 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.446069 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:17.446098 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.446082 2569 projected.go:194] Error preparing data for projected volume kube-api-access-rh6ms for pod openshift-network-diagnostics/network-check-target-2cw78: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:17.446257 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.446145 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms podName:e338cd11-dfc1-43bb-8092-32e10f1e62a5 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:25.446126315 +0000 UTC m=+18.287163511 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-rh6ms" (UniqueName: "kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms") pod "network-check-target-2cw78" (UID: "e338cd11-dfc1-43bb-8092-32e10f1e62a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:17.707589 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:17.706863 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:17.707589 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.706997 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:17.707589 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:17.707436 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:17.707589 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:17.707548 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:19.705935 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:19.705873 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:19.706339 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:19.705877 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:19.706339 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:19.705964 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:19.706339 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:19.706077 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:21.706076 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:21.706045 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:21.706480 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:21.706052 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:21.706480 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:21.706169 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:21.706480 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:21.706249 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:23.702600 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:23.702564 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:23.703085 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:23.702570 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:23.703085 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:23.702700 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:23.703085 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:23.702796 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:25.303976 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:25.303899 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:25.304419 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.304039 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:25.304419 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.304118 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:21:41.304096154 +0000 UTC m=+34.145133359 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:25.506010 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:25.505980 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:25.506157 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.506092 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:25.506157 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.506105 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:25.506157 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.506114 2569 projected.go:194] Error preparing data for projected volume kube-api-access-rh6ms for pod openshift-network-diagnostics/network-check-target-2cw78: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:25.506255 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.506158 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms podName:e338cd11-dfc1-43bb-8092-32e10f1e62a5 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:41.506145145 +0000 UTC m=+34.347182335 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-rh6ms" (UniqueName: "kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms") pod "network-check-target-2cw78" (UID: "e338cd11-dfc1-43bb-8092-32e10f1e62a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:25.702668 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:25.702588 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:25.702668 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:25.702638 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:25.702943 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.702730 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:25.702943 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:25.702866 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:27.706201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:27.705560 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:27.706201 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:27.705863 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:27.706201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:27.705976 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:27.706201 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:27.706055 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:27.784844 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:27.784792 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-b65w4" event={"ID":"d7b21dc1-e456-4f3a-827a-6d17899801ad","Type":"ContainerStarted","Data":"bd6774892723e621820c50e0bb13585905f5f2b6810566e9594b23d68cd92474"} Apr 17 09:21:27.792391 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:27.792344 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" event={"ID":"904ead54-94c1-4e84-9851-0c8506ed7bbc","Type":"ContainerStarted","Data":"6215c00c65bc16815c7387f7d5cc91e425bef4a963a2ed3abfbb2381056fefbe"} Apr 17 09:21:27.792951 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:27.792921 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:27.799676 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:27.799262 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-b65w4" podStartSLOduration=11.861775571 podStartE2EDuration="20.799245831s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.405072659 +0000 UTC m=+3.246109852" lastFinishedPulling="2026-04-17 09:21:19.342542917 +0000 UTC m=+12.183580112" observedRunningTime="2026-04-17 09:21:27.798520239 +0000 UTC m=+20.639557451" watchObservedRunningTime="2026-04-17 09:21:27.799245831 +0000 UTC m=+20.640283045" Apr 17 09:21:27.819270 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:27.819217 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" podStartSLOduration=3.671327662 podStartE2EDuration="20.819197528s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.405197385 +0000 UTC m=+3.246234576" lastFinishedPulling="2026-04-17 09:21:27.553067249 +0000 UTC m=+20.394104442" observedRunningTime="2026-04-17 09:21:27.818660423 +0000 UTC m=+20.659697642" watchObservedRunningTime="2026-04-17 09:21:27.819197528 +0000 UTC m=+20.660234741" Apr 17 09:21:28.803191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.802987 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-t5ckx" event={"ID":"09912815-97d7-47ef-939f-9f3815c992c1","Type":"ContainerStarted","Data":"8e588e974895c8d9728e28c5bf2c5dfea94e9505a3f8a4d03792aa1092d43f94"} Apr 17 09:21:28.804264 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.804230 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-rjm7r" event={"ID":"0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32","Type":"ContainerStarted","Data":"688aeb4e2fd63672c287ac3601eaed3eed253b82f13eb405099d798126d26968"} Apr 17 09:21:28.806446 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806421 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/0.log" Apr 17 09:21:28.806693 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806674 2569 generic.go:358] "Generic (PLEG): container finished" podID="1ea35608-b7a5-4952-8b6c-f24a85b0ff89" containerID="a8e9e952cc78a5ce6fb96ea2199f719ff7fd05bcbb33f17bb092d129896cfa71" exitCode=1 Apr 17 09:21:28.806773 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806729 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"b7113990e5e4e68f9b7af6c0f01461bd83750314ba101ba43ce69c3d45866a49"} Apr 17 09:21:28.806773 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806746 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"319461ce7665875db3a70ac21fdc4489b88def467cf127da4e98b6e33fbbe1ca"} Apr 17 09:21:28.806773 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806755 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"1ce354a83d5c8b9d12d80eafb53a80a4d1e0c642add962f294aba02c5c8cbb16"} Apr 17 09:21:28.806773 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806764 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"4fbb9de734035055be8fa98263d9708a876df10f410f92cb267a81477fec4711"} Apr 17 09:21:28.806773 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806772 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerDied","Data":"a8e9e952cc78a5ce6fb96ea2199f719ff7fd05bcbb33f17bb092d129896cfa71"} Apr 17 09:21:28.806983 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.806782 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"b37c5ef42573df7eba856fb044527a057bd6415e15f8be7c319f3e5f6aabc79b"} Apr 17 09:21:28.807921 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.807894 2569 generic.go:358] "Generic (PLEG): container finished" podID="70fa4740-d7b3-4bc0-8829-d52ad2ac1144" containerID="deefbe54e89230973052b91b81f17b0a50e207a9a12c06a032051d110e2eacb6" exitCode=0 Apr 17 09:21:28.807996 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.807961 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerDied","Data":"deefbe54e89230973052b91b81f17b0a50e207a9a12c06a032051d110e2eacb6"} Apr 17 09:21:28.809124 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.809101 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-b5dkm" event={"ID":"9fcd330e-f463-420a-a91c-f2a6e23daebc","Type":"ContainerStarted","Data":"7c71767a39fa97eef7c59012b0d60fd1d8d0a57374e397f2734d34d6332accbf"} Apr 17 09:21:28.810394 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.810372 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9bwlv" event={"ID":"e686455d-e53d-4a81-b3c4-74eb05f3da00","Type":"ContainerStarted","Data":"c986ed28477b074e1c9277c6491784772e8013db9f9f8600e0f31f07022831f1"} Apr 17 09:21:28.816389 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.816356 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-t5ckx" podStartSLOduration=4.6914005880000005 podStartE2EDuration="21.816346684s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.398342715 +0000 UTC m=+3.239379916" lastFinishedPulling="2026-04-17 09:21:27.523288811 +0000 UTC m=+20.364326012" observedRunningTime="2026-04-17 09:21:28.816203889 +0000 UTC m=+21.657241100" watchObservedRunningTime="2026-04-17 09:21:28.816346684 +0000 UTC m=+21.657383896" Apr 17 09:21:28.850508 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.850468 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-b5dkm" podStartSLOduration=4.702172607 podStartE2EDuration="21.85045702s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.374997973 +0000 UTC m=+3.216035165" lastFinishedPulling="2026-04-17 09:21:27.523282387 +0000 UTC m=+20.364319578" observedRunningTime="2026-04-17 09:21:28.85032043 +0000 UTC m=+21.691357642" watchObservedRunningTime="2026-04-17 09:21:28.85045702 +0000 UTC m=+21.691494233" Apr 17 09:21:28.867916 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:28.867874 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-9bwlv" podStartSLOduration=4.687757514 podStartE2EDuration="21.867862326s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.373292972 +0000 UTC m=+3.214330165" lastFinishedPulling="2026-04-17 09:21:27.553397787 +0000 UTC m=+20.394434977" observedRunningTime="2026-04-17 09:21:28.867756697 +0000 UTC m=+21.708793908" watchObservedRunningTime="2026-04-17 09:21:28.867862326 +0000 UTC m=+21.708899538" Apr 17 09:21:29.173135 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.173104 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:29.173856 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.173835 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:29.250752 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.250730 2569 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 17 09:21:29.644951 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.644421 2569 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-17T09:21:29.250748218Z","UUID":"06f9f337-4dca-4ee9-9cda-20c3051633b9","Handler":null,"Name":"","Endpoint":""} Apr 17 09:21:29.646873 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.646849 2569 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 17 09:21:29.647001 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.646882 2569 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 17 09:21:29.702778 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.702523 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:29.702778 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:29.702644 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:29.702778 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.702727 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:29.703070 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:29.702840 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:29.813665 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.813579 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" event={"ID":"904ead54-94c1-4e84-9851-0c8506ed7bbc","Type":"ContainerStarted","Data":"38bc509721dc944969c4bf51e57e2c97e001e3df7ea031a83adc3b79b6f23d7c"} Apr 17 09:21:29.815042 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.815000 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-dd7tv" event={"ID":"2f0ca441-19bb-4396-adb5-c2b223622c08","Type":"ContainerStarted","Data":"36de61ac63906e4a350462c6d69e49cdb6cb2a0886790d1e919017b258f40d45"} Apr 17 09:21:29.816108 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.816083 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-b65w4" Apr 17 09:21:29.828939 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:29.828900 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-dd7tv" podStartSLOduration=5.704003075 podStartE2EDuration="22.828888102s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.398334682 +0000 UTC m=+3.239371871" lastFinishedPulling="2026-04-17 09:21:27.523219699 +0000 UTC m=+20.364256898" observedRunningTime="2026-04-17 09:21:29.828871397 +0000 UTC m=+22.669908610" watchObservedRunningTime="2026-04-17 09:21:29.828888102 +0000 UTC m=+22.669925307" Apr 17 09:21:30.820049 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:30.819841 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/0.log" Apr 17 09:21:30.820575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:30.820398 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"cec6be2192a6d7c90b1f1636acdc21ae4cbe1e1e2278347310a256d9128619e2"} Apr 17 09:21:30.822434 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:30.822402 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" event={"ID":"904ead54-94c1-4e84-9851-0c8506ed7bbc","Type":"ContainerStarted","Data":"0ffa5af64b9f6c47c0fd2fd1e74cf9312641551d4dc7d297ffb24a2b2dfffcfe"} Apr 17 09:21:30.838860 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:30.838792 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-8l4wz" podStartSLOduration=3.810345433 podStartE2EDuration="23.838774353s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.376722298 +0000 UTC m=+3.217759489" lastFinishedPulling="2026-04-17 09:21:30.405151207 +0000 UTC m=+23.246188409" observedRunningTime="2026-04-17 09:21:30.838281437 +0000 UTC m=+23.679318654" watchObservedRunningTime="2026-04-17 09:21:30.838774353 +0000 UTC m=+23.679811588" Apr 17 09:21:31.703013 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:31.702980 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:31.703188 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:31.703100 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:31.703188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:31.703148 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:31.703301 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:31.703251 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:33.702448 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.702265 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:33.703136 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.702280 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:33.703136 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:33.702524 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:33.703136 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:33.702636 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:33.830131 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.830109 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/0.log" Apr 17 09:21:33.830471 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.830446 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"458c2f5a277d24cc1ff8e3938657609d821351b0b0dcabe273ebf0051d2c1bc7"} Apr 17 09:21:33.830749 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.830734 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:33.830946 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.830930 2569 scope.go:117] "RemoveContainer" containerID="a8e9e952cc78a5ce6fb96ea2199f719ff7fd05bcbb33f17bb092d129896cfa71" Apr 17 09:21:33.832013 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.831942 2569 generic.go:358] "Generic (PLEG): container finished" podID="70fa4740-d7b3-4bc0-8829-d52ad2ac1144" containerID="f6af88b0cc8ffaa167d7f5f843efd8a5b092d0924a2076a09f0bbbfcb827540a" exitCode=0 Apr 17 09:21:33.832013 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.831997 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerDied","Data":"f6af88b0cc8ffaa167d7f5f843efd8a5b092d0924a2076a09f0bbbfcb827540a"} Apr 17 09:21:33.845979 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:33.845961 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:34.786278 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.786069 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-2cw78"] Apr 17 09:21:34.786698 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.786367 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:34.786698 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:34.786472 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:34.796987 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.796958 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4kbg5"] Apr 17 09:21:34.797107 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.797079 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:34.797209 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:34.797187 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:34.837403 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.837346 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/0.log" Apr 17 09:21:34.837687 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.837656 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" event={"ID":"1ea35608-b7a5-4952-8b6c-f24a85b0ff89","Type":"ContainerStarted","Data":"723949ed2088b06c9fff640c8468d0ea78cbf9ecc9eb9ba245b9edb61d0cc40b"} Apr 17 09:21:34.837968 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.837941 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:34.837968 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.837971 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:34.854036 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.854014 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:21:34.862694 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:34.862656 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" podStartSLOduration=10.657718604 podStartE2EDuration="27.862645133s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.404993644 +0000 UTC m=+3.246030837" lastFinishedPulling="2026-04-17 09:21:27.609920168 +0000 UTC m=+20.450957366" observedRunningTime="2026-04-17 09:21:34.861202048 +0000 UTC m=+27.702239260" watchObservedRunningTime="2026-04-17 09:21:34.862645133 +0000 UTC m=+27.703682342" Apr 17 09:21:35.841218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:35.841183 2569 generic.go:358] "Generic (PLEG): container finished" podID="70fa4740-d7b3-4bc0-8829-d52ad2ac1144" containerID="f6e559cbcf14947bb4abf1d6e7fe3f01ff6572b6818e3dbc4136a16edc2d8de5" exitCode=0 Apr 17 09:21:35.841218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:35.841212 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerDied","Data":"f6e559cbcf14947bb4abf1d6e7fe3f01ff6572b6818e3dbc4136a16edc2d8de5"} Apr 17 09:21:36.702192 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:36.702161 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:36.702192 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:36.702187 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:36.702368 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:36.702257 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:36.702368 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:36.702322 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:37.846115 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:37.846083 2569 generic.go:358] "Generic (PLEG): container finished" podID="70fa4740-d7b3-4bc0-8829-d52ad2ac1144" containerID="0cc4721d89f12a803ad9c1ef6f564f23745c155a8b8152d043e24e4a2814de27" exitCode=0 Apr 17 09:21:37.846674 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:37.846134 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerDied","Data":"0cc4721d89f12a803ad9c1ef6f564f23745c155a8b8152d043e24e4a2814de27"} Apr 17 09:21:38.702638 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:38.702603 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:38.702832 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:38.702646 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:38.702832 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:38.702748 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:38.702951 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:38.702897 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:40.702812 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:40.702779 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:40.703277 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:40.702779 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:40.703277 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:40.702923 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4kbg5" podUID="833e0b85-0aba-4f60-bbf6-a215ed02831a" Apr 17 09:21:40.703277 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:40.703050 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2cw78" podUID="e338cd11-dfc1-43bb-8092-32e10f1e62a5" Apr 17 09:21:40.970962 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:40.970885 2569 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-152.ec2.internal" event="NodeReady" Apr 17 09:21:40.971132 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:40.971047 2569 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 17 09:21:41.012329 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.012294 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-27fm4"] Apr 17 09:21:41.030482 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.030448 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pf795"] Apr 17 09:21:41.030644 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.030618 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.033067 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.033043 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-nblxl\"" Apr 17 09:21:41.033209 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.033082 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 17 09:21:41.033339 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.033323 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 17 09:21:41.042286 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.042266 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-27fm4"] Apr 17 09:21:41.042286 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.042289 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pf795"] Apr 17 09:21:41.042451 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.042386 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:41.044584 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.044560 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 17 09:21:41.044723 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.044708 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 17 09:21:41.044878 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.044860 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9szrj\"" Apr 17 09:21:41.044952 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.044860 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 17 09:21:41.123095 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.123053 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdbw7\" (UniqueName: \"kubernetes.io/projected/b171f472-b404-4138-bf91-727a857e3224-kube-api-access-zdbw7\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.123252 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.123165 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:41.123252 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.123194 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/b171f472-b404-4138-bf91-727a857e3224-tmp-dir\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.123252 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.123224 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndlkl\" (UniqueName: \"kubernetes.io/projected/e5c6f5f4-e252-4726-9992-25c01cae77a3-kube-api-access-ndlkl\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:41.123404 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.123250 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b171f472-b404-4138-bf91-727a857e3224-config-volume\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.123404 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.123365 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.224047 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.223970 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:41.224047 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.224006 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/b171f472-b404-4138-bf91-727a857e3224-tmp-dir\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.224047 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.224036 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ndlkl\" (UniqueName: \"kubernetes.io/projected/e5c6f5f4-e252-4726-9992-25c01cae77a3-kube-api-access-ndlkl\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:41.224326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.224062 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b171f472-b404-4138-bf91-727a857e3224-config-volume\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.224326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.224124 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.224326 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.224148 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:41.224326 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.224222 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:41.724201073 +0000 UTC m=+34.565238266 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:21:41.224326 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.224240 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:41.224326 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.224304 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:41.724286863 +0000 UTC m=+34.565324053 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:21:41.224326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.224152 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zdbw7\" (UniqueName: \"kubernetes.io/projected/b171f472-b404-4138-bf91-727a857e3224-kube-api-access-zdbw7\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.224621 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.224391 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/b171f472-b404-4138-bf91-727a857e3224-tmp-dir\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.224720 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.224693 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b171f472-b404-4138-bf91-727a857e3224-config-volume\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.234579 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.234450 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdbw7\" (UniqueName: \"kubernetes.io/projected/b171f472-b404-4138-bf91-727a857e3224-kube-api-access-zdbw7\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.234723 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.234709 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndlkl\" (UniqueName: \"kubernetes.io/projected/e5c6f5f4-e252-4726-9992-25c01cae77a3-kube-api-access-ndlkl\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:41.325200 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.325156 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:41.325371 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.325278 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:41.325371 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.325361 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:22:13.325339083 +0000 UTC m=+66.166376287 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:41.526144 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.526110 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:41.526363 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.526261 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:41.526363 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.526283 2569 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:41.526363 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.526294 2569 projected.go:194] Error preparing data for projected volume kube-api-access-rh6ms for pod openshift-network-diagnostics/network-check-target-2cw78: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:41.526535 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.526364 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms podName:e338cd11-dfc1-43bb-8092-32e10f1e62a5 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:13.526342562 +0000 UTC m=+66.367379756 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-rh6ms" (UniqueName: "kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms") pod "network-check-target-2cw78" (UID: "e338cd11-dfc1-43bb-8092-32e10f1e62a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:41.727322 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.727286 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:41.727769 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:41.727356 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:41.727769 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.727445 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:41.727769 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.727478 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:41.727769 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.727522 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:42.727500324 +0000 UTC m=+35.568537514 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:21:41.727769 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:41.727554 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:42.727537842 +0000 UTC m=+35.568575037 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:21:42.702766 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.702727 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:21:42.703090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.702960 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:21:42.706963 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.706891 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-8wjkc\"" Apr 17 09:21:42.706963 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.706891 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 17 09:21:42.707146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.706985 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-4w67p\"" Apr 17 09:21:42.707146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.706998 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 17 09:21:42.707146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.706891 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 17 09:21:42.736140 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.736118 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:42.736484 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:42.736165 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:42.736484 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:42.736265 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:42.736484 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:42.736286 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:42.736484 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:42.736317 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:44.736299389 +0000 UTC m=+37.577336581 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:21:42.736484 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:42.736347 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:44.736327312 +0000 UTC m=+37.577364516 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:21:44.753550 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:44.753474 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:44.753550 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:44.753531 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:44.753985 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:44.753621 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:44.753985 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:44.753633 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:44.753985 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:44.753669 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:48.753654529 +0000 UTC m=+41.594691718 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:21:44.753985 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:44.753697 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:48.753677294 +0000 UTC m=+41.594714485 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:21:44.861072 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:44.861040 2569 generic.go:358] "Generic (PLEG): container finished" podID="70fa4740-d7b3-4bc0-8829-d52ad2ac1144" containerID="b98185b604a092258958e0710279add7fd0251cfb69ee5c272c421c8ff6efbd6" exitCode=0 Apr 17 09:21:44.861241 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:44.861087 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerDied","Data":"b98185b604a092258958e0710279add7fd0251cfb69ee5c272c421c8ff6efbd6"} Apr 17 09:21:45.865787 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:45.865755 2569 generic.go:358] "Generic (PLEG): container finished" podID="70fa4740-d7b3-4bc0-8829-d52ad2ac1144" containerID="1cc943a1b582dd9c2857a8346a43ba37c366468adc936609f6ced81291514cbe" exitCode=0 Apr 17 09:21:45.866187 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:45.865806 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerDied","Data":"1cc943a1b582dd9c2857a8346a43ba37c366468adc936609f6ced81291514cbe"} Apr 17 09:21:46.870786 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:46.870752 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" event={"ID":"70fa4740-d7b3-4bc0-8829-d52ad2ac1144","Type":"ContainerStarted","Data":"78e25fc617ff22156f2906f30840c39c0ff3a2df31846821e19e37068a9109a8"} Apr 17 09:21:46.892945 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:46.892904 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-g7bkg" podStartSLOduration=6.521919166 podStartE2EDuration="39.892889564s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:21:10.401500707 +0000 UTC m=+3.242537912" lastFinishedPulling="2026-04-17 09:21:43.772471116 +0000 UTC m=+36.613508310" observedRunningTime="2026-04-17 09:21:46.891501024 +0000 UTC m=+39.732538235" watchObservedRunningTime="2026-04-17 09:21:46.892889564 +0000 UTC m=+39.733926776" Apr 17 09:21:48.781190 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:48.781152 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:48.781686 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:48.781240 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:48.781686 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:48.781287 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:48.781686 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:48.781349 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:56.781333468 +0000 UTC m=+49.622370658 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:21:48.781686 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:48.781351 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:48.781686 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:48.781401 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:56.781386092 +0000 UTC m=+49.622423303 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:21:56.833015 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:56.832983 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:21:56.833380 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:21:56.833040 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:21:56.833380 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:56.833128 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:56.833380 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:56.833136 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:56.833380 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:56.833184 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:12.83317153 +0000 UTC m=+65.674208720 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:21:56.833380 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:21:56.833200 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:12.833192429 +0000 UTC m=+65.674229619 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:22:06.856577 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:06.856544 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd2rv" Apr 17 09:22:12.838101 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:12.838066 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:22:12.838500 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:12.838126 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:22:12.838500 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:12.838214 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:22:12.838500 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:12.838289 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:44.838273088 +0000 UTC m=+97.679310277 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:22:12.838500 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:12.838218 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:22:12.838500 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:12.838352 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:44.838338581 +0000 UTC m=+97.679375770 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:22:13.341280 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.341246 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:22:13.343868 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.343842 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 17 09:22:13.351655 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:13.351632 2569 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 17 09:22:13.351763 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:13.351734 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs podName:833e0b85-0aba-4f60-bbf6-a215ed02831a nodeName:}" failed. No retries permitted until 2026-04-17 09:23:17.351707884 +0000 UTC m=+130.192745078 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs") pod "network-metrics-daemon-4kbg5" (UID: "833e0b85-0aba-4f60-bbf6-a215ed02831a") : secret "metrics-daemon-secret" not found Apr 17 09:22:13.543147 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.543121 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:22:13.545645 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.545629 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 17 09:22:13.555367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.555349 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 17 09:22:13.567423 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.567400 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh6ms\" (UniqueName: \"kubernetes.io/projected/e338cd11-dfc1-43bb-8092-32e10f1e62a5-kube-api-access-rh6ms\") pod \"network-check-target-2cw78\" (UID: \"e338cd11-dfc1-43bb-8092-32e10f1e62a5\") " pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:22:13.616634 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.616584 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-4w67p\"" Apr 17 09:22:13.624423 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.624403 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:22:13.742008 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.741981 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-2cw78"] Apr 17 09:22:13.745572 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:22:13.745543 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode338cd11_dfc1_43bb_8092_32e10f1e62a5.slice/crio-2a21682eadd2f5f955803f28e40251e0d2f1eed95dde5102f81e732cc16b0d33 WatchSource:0}: Error finding container 2a21682eadd2f5f955803f28e40251e0d2f1eed95dde5102f81e732cc16b0d33: Status 404 returned error can't find the container with id 2a21682eadd2f5f955803f28e40251e0d2f1eed95dde5102f81e732cc16b0d33 Apr 17 09:22:13.919183 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:13.919122 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-2cw78" event={"ID":"e338cd11-dfc1-43bb-8092-32e10f1e62a5","Type":"ContainerStarted","Data":"2a21682eadd2f5f955803f28e40251e0d2f1eed95dde5102f81e732cc16b0d33"} Apr 17 09:22:16.926292 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:16.926258 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-2cw78" event={"ID":"e338cd11-dfc1-43bb-8092-32e10f1e62a5","Type":"ContainerStarted","Data":"a93a27ffae0c03f543ac575460dadfa4208777452c373a90c80d19ee90045bca"} Apr 17 09:22:16.926610 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:16.926391 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:22:16.941177 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:16.941048 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-2cw78" podStartSLOduration=67.397766128 podStartE2EDuration="1m9.941034794s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:22:13.747282911 +0000 UTC m=+66.588320100" lastFinishedPulling="2026-04-17 09:22:16.290551573 +0000 UTC m=+69.131588766" observedRunningTime="2026-04-17 09:22:16.940863207 +0000 UTC m=+69.781900422" watchObservedRunningTime="2026-04-17 09:22:16.941034794 +0000 UTC m=+69.782072007" Apr 17 09:22:44.938678 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:44.938650 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:22:44.939094 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:44.938710 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:22:44.939094 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:44.938799 2569 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:22:44.939094 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:44.938872 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert podName:e5c6f5f4-e252-4726-9992-25c01cae77a3 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:48.938855685 +0000 UTC m=+161.779892888 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert") pod "ingress-canary-pf795" (UID: "e5c6f5f4-e252-4726-9992-25c01cae77a3") : secret "canary-serving-cert" not found Apr 17 09:22:44.939094 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:44.938799 2569 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:22:44.939094 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:44.938945 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls podName:b171f472-b404-4138-bf91-727a857e3224 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:48.938932713 +0000 UTC m=+161.779969903 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls") pod "dns-default-27fm4" (UID: "b171f472-b404-4138-bf91-727a857e3224") : secret "dns-default-metrics-tls" not found Apr 17 09:22:47.930414 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:47.930383 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-2cw78" Apr 17 09:22:55.964811 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.964777 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft"] Apr 17 09:22:55.969517 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.969492 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-z9nkj"] Apr 17 09:22:55.969668 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.969649 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:55.972466 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.972445 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:55.972584 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.972508 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Apr 17 09:22:55.972584 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.972565 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Apr 17 09:22:55.973202 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.973180 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-6vwhv\"" Apr 17 09:22:55.973308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.973228 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:22:55.974455 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.974429 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:22:55.974589 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.974572 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Apr 17 09:22:55.975468 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.975452 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Apr 17 09:22:55.975562 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.975474 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-l4kf2\"" Apr 17 09:22:55.975562 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.975497 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Apr 17 09:22:55.979809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.979789 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Apr 17 09:22:55.980050 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.980027 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft"] Apr 17 09:22:55.984391 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:55.984368 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-z9nkj"] Apr 17 09:22:56.011584 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.011544 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q64wd\" (UniqueName: \"kubernetes.io/projected/8f81e523-bc6a-40b3-8e85-891b04d3707b-kube-api-access-q64wd\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:56.011766 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.011615 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-serving-cert\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.011766 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.011642 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-trusted-ca\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.011766 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.011671 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx6k7\" (UniqueName: \"kubernetes.io/projected/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-kube-api-access-mx6k7\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.011766 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.011699 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-config\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.011943 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.011787 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:56.112128 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.112095 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:56.112317 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.112135 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-q64wd\" (UniqueName: \"kubernetes.io/projected/8f81e523-bc6a-40b3-8e85-891b04d3707b-kube-api-access-q64wd\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:56.112317 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.112187 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-serving-cert\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.112317 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.112210 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-trusted-ca\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.112317 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.112242 2569 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 09:22:56.112317 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.112258 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mx6k7\" (UniqueName: \"kubernetes.io/projected/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-kube-api-access-mx6k7\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.112317 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.112290 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-config\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.112317 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.112304 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls podName:8f81e523-bc6a-40b3-8e85-891b04d3707b nodeName:}" failed. No retries permitted until 2026-04-17 09:22:56.612285309 +0000 UTC m=+109.453322518 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-pdqft" (UID: "8f81e523-bc6a-40b3-8e85-891b04d3707b") : secret "samples-operator-tls" not found Apr 17 09:22:56.112919 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.112893 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-config\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.113579 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.113556 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-trusted-ca\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.118050 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.118026 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-serving-cert\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.121989 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.121969 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj"] Apr 17 09:22:56.125015 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.124998 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl"] Apr 17 09:22:56.125143 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.125127 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" Apr 17 09:22:56.127811 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.127772 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.133181 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.133162 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-75d68f6767-9txgs"] Apr 17 09:22:56.133921 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.133901 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"network-diagnostics-dockercfg-v2446\"" Apr 17 09:22:56.134139 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.134126 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Apr 17 09:22:56.134375 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.134359 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:22:56.134430 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.134389 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Apr 17 09:22:56.134711 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.134695 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-kptsc\"" Apr 17 09:22:56.136206 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.136186 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.137690 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.137668 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl"] Apr 17 09:22:56.138695 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.138673 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx6k7\" (UniqueName: \"kubernetes.io/projected/ef4e2c76-47f7-42fb-a0e4-d95b609d0101-kube-api-access-mx6k7\") pod \"console-operator-9d4b6777b-z9nkj\" (UID: \"ef4e2c76-47f7-42fb-a0e4-d95b609d0101\") " pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.140002 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.139985 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Apr 17 09:22:56.142357 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.142337 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 17 09:22:56.142875 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.142597 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 17 09:22:56.142875 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.142686 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-brmrj\"" Apr 17 09:22:56.142875 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.142771 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 17 09:22:56.143569 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.143546 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj"] Apr 17 09:22:56.152186 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.152168 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 17 09:22:56.172461 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.170349 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-75d68f6767-9txgs"] Apr 17 09:22:56.174424 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.174401 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-q64wd\" (UniqueName: \"kubernetes.io/projected/8f81e523-bc6a-40b3-8e85-891b04d3707b-kube-api-access-q64wd\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:56.213238 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213212 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tqrp\" (UniqueName: \"kubernetes.io/projected/89378780-0d1b-40cd-930c-da0d27751523-kube-api-access-7tqrp\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.213379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213244 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/414546fe-e0a7-4aad-bf7e-fa775be0036d-ca-trust-extracted\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.213379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213274 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89378780-0d1b-40cd-930c-da0d27751523-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.213379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213289 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgwjd\" (UniqueName: \"kubernetes.io/projected/abd1dcbf-ede8-4fa7-be3e-02a819a3421c-kube-api-access-wgwjd\") pod \"network-check-source-8894fc9bd-pzchj\" (UID: \"abd1dcbf-ede8-4fa7-be3e-02a819a3421c\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" Apr 17 09:22:56.213379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213331 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-certificates\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.213379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213346 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-bound-sa-token\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.213598 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213411 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.213598 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213481 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-trusted-ca\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.213598 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213526 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-installation-pull-secrets\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.213598 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213543 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vt52\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-kube-api-access-7vt52\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.213598 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213567 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89378780-0d1b-40cd-930c-da0d27751523-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.213598 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.213589 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-image-registry-private-configuration\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.214672 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.214655 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg"] Apr 17 09:22:56.217827 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.217779 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.221232 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.221213 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-4xs4t\"" Apr 17 09:22:56.221464 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.221450 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-tls\"" Apr 17 09:22:56.222411 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.222393 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 17 09:22:56.224971 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.224953 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 17 09:22:56.225999 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.225985 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemetry-config\"" Apr 17 09:22:56.237474 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.237445 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg"] Apr 17 09:22:56.286247 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.286227 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:22:56.314193 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314150 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7tqrp\" (UniqueName: \"kubernetes.io/projected/89378780-0d1b-40cd-930c-da0d27751523-kube-api-access-7tqrp\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.314343 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314207 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/414546fe-e0a7-4aad-bf7e-fa775be0036d-ca-trust-extracted\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.314391 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314369 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89378780-0d1b-40cd-930c-da0d27751523-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.314439 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314400 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wgwjd\" (UniqueName: \"kubernetes.io/projected/abd1dcbf-ede8-4fa7-be3e-02a819a3421c-kube-api-access-wgwjd\") pod \"network-check-source-8894fc9bd-pzchj\" (UID: \"abd1dcbf-ede8-4fa7-be3e-02a819a3421c\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" Apr 17 09:22:56.314439 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314430 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmg4v\" (UniqueName: \"kubernetes.io/projected/22a71e37-6733-4843-a7c4-2027bbd017e5-kube-api-access-lmg4v\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.314532 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314463 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-certificates\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.314532 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314489 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-bound-sa-token\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.314652 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314629 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.314838 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314670 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/22a71e37-6733-4843-a7c4-2027bbd017e5-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.314838 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314700 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.314838 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.314716 2569 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:22:56.314838 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.314736 2569 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-75d68f6767-9txgs: secret "image-registry-tls" not found Apr 17 09:22:56.314838 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314751 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-trusted-ca\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.314838 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314631 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/414546fe-e0a7-4aad-bf7e-fa775be0036d-ca-trust-extracted\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.314838 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.314806 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls podName:414546fe-e0a7-4aad-bf7e-fa775be0036d nodeName:}" failed. No retries permitted until 2026-04-17 09:22:56.814786047 +0000 UTC m=+109.655823238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls") pod "image-registry-75d68f6767-9txgs" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d") : secret "image-registry-tls" not found Apr 17 09:22:56.315241 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314886 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-installation-pull-secrets\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.315241 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314920 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7vt52\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-kube-api-access-7vt52\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.315241 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314954 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89378780-0d1b-40cd-930c-da0d27751523-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.315241 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.314973 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-image-registry-private-configuration\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.315241 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.315094 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-certificates\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.315626 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.315606 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89378780-0d1b-40cd-930c-da0d27751523-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.316329 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.316307 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-trusted-ca\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.317055 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.317031 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89378780-0d1b-40cd-930c-da0d27751523-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.317676 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.317659 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-image-registry-private-configuration\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.317853 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.317807 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-installation-pull-secrets\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.323731 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.323709 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tqrp\" (UniqueName: \"kubernetes.io/projected/89378780-0d1b-40cd-930c-da0d27751523-kube-api-access-7tqrp\") pod \"kube-storage-version-migrator-operator-6769c5d45-zj5rl\" (UID: \"89378780-0d1b-40cd-930c-da0d27751523\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.323963 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.323943 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vt52\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-kube-api-access-7vt52\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.324058 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.323973 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgwjd\" (UniqueName: \"kubernetes.io/projected/abd1dcbf-ede8-4fa7-be3e-02a819a3421c-kube-api-access-wgwjd\") pod \"network-check-source-8894fc9bd-pzchj\" (UID: \"abd1dcbf-ede8-4fa7-be3e-02a819a3421c\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" Apr 17 09:22:56.324058 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.323943 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-bound-sa-token\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.413897 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.413865 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-z9nkj"] Apr 17 09:22:56.415946 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.415925 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lmg4v\" (UniqueName: \"kubernetes.io/projected/22a71e37-6733-4843-a7c4-2027bbd017e5-kube-api-access-lmg4v\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.416017 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.415978 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/22a71e37-6733-4843-a7c4-2027bbd017e5-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.416101 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.416086 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.416281 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.416261 2569 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 09:22:56.416364 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.416349 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls podName:22a71e37-6733-4843-a7c4-2027bbd017e5 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:56.916328055 +0000 UTC m=+109.757365246 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-9rpfg" (UID: "22a71e37-6733-4843-a7c4-2027bbd017e5") : secret "cluster-monitoring-operator-tls" not found Apr 17 09:22:56.416639 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:22:56.416618 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef4e2c76_47f7_42fb_a0e4_d95b609d0101.slice/crio-f725ea94bae28d7c9aab86a0c4e7f7dea248fce06c208a4d082fad2351a6c3f3 WatchSource:0}: Error finding container f725ea94bae28d7c9aab86a0c4e7f7dea248fce06c208a4d082fad2351a6c3f3: Status 404 returned error can't find the container with id f725ea94bae28d7c9aab86a0c4e7f7dea248fce06c208a4d082fad2351a6c3f3 Apr 17 09:22:56.416739 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.416722 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/22a71e37-6733-4843-a7c4-2027bbd017e5-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.425774 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.425753 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmg4v\" (UniqueName: \"kubernetes.io/projected/22a71e37-6733-4843-a7c4-2027bbd017e5-kube-api-access-lmg4v\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.439093 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.434169 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" Apr 17 09:22:56.445695 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.445671 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" Apr 17 09:22:56.548908 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.548878 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj"] Apr 17 09:22:56.551528 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:22:56.551499 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabd1dcbf_ede8_4fa7_be3e_02a819a3421c.slice/crio-6fa4920ef11fad5da5a868c88921016948b36bfd622713f00ddb3ae05b10782f WatchSource:0}: Error finding container 6fa4920ef11fad5da5a868c88921016948b36bfd622713f00ddb3ae05b10782f: Status 404 returned error can't find the container with id 6fa4920ef11fad5da5a868c88921016948b36bfd622713f00ddb3ae05b10782f Apr 17 09:22:56.562559 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.562533 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl"] Apr 17 09:22:56.566300 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:22:56.566273 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89378780_0d1b_40cd_930c_da0d27751523.slice/crio-fcf2c800afb61b7b5c79717c7bf66f94829502f2c01367204727a3a1d39937a9 WatchSource:0}: Error finding container fcf2c800afb61b7b5c79717c7bf66f94829502f2c01367204727a3a1d39937a9: Status 404 returned error can't find the container with id fcf2c800afb61b7b5c79717c7bf66f94829502f2c01367204727a3a1d39937a9 Apr 17 09:22:56.618359 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.618326 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:56.618457 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.618449 2569 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 09:22:56.618501 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.618493 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls podName:8f81e523-bc6a-40b3-8e85-891b04d3707b nodeName:}" failed. No retries permitted until 2026-04-17 09:22:57.618479877 +0000 UTC m=+110.459517067 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-pdqft" (UID: "8f81e523-bc6a-40b3-8e85-891b04d3707b") : secret "samples-operator-tls" not found Apr 17 09:22:56.819012 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.818972 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:56.819192 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.819143 2569 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:22:56.819192 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.819165 2569 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-75d68f6767-9txgs: secret "image-registry-tls" not found Apr 17 09:22:56.819307 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.819231 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls podName:414546fe-e0a7-4aad-bf7e-fa775be0036d nodeName:}" failed. No retries permitted until 2026-04-17 09:22:57.819210588 +0000 UTC m=+110.660247778 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls") pod "image-registry-75d68f6767-9txgs" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d") : secret "image-registry-tls" not found Apr 17 09:22:56.919689 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:56.919655 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:56.919856 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.919840 2569 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 09:22:56.919932 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:56.919920 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls podName:22a71e37-6733-4843-a7c4-2027bbd017e5 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:57.919899751 +0000 UTC m=+110.760936957 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-9rpfg" (UID: "22a71e37-6733-4843-a7c4-2027bbd017e5") : secret "cluster-monitoring-operator-tls" not found Apr 17 09:22:57.000689 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.000652 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" event={"ID":"abd1dcbf-ede8-4fa7-be3e-02a819a3421c","Type":"ContainerStarted","Data":"baf8e4eca5de5e563a2910dfd5869eb6d0c6e69860ddfce8fdc677dfaf10f543"} Apr 17 09:22:57.001147 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.000699 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" event={"ID":"abd1dcbf-ede8-4fa7-be3e-02a819a3421c","Type":"ContainerStarted","Data":"6fa4920ef11fad5da5a868c88921016948b36bfd622713f00ddb3ae05b10782f"} Apr 17 09:22:57.001906 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.001863 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" event={"ID":"89378780-0d1b-40cd-930c-da0d27751523","Type":"ContainerStarted","Data":"fcf2c800afb61b7b5c79717c7bf66f94829502f2c01367204727a3a1d39937a9"} Apr 17 09:22:57.002869 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.002846 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" event={"ID":"ef4e2c76-47f7-42fb-a0e4-d95b609d0101","Type":"ContainerStarted","Data":"f725ea94bae28d7c9aab86a0c4e7f7dea248fce06c208a4d082fad2351a6c3f3"} Apr 17 09:22:57.018655 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.018606 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-pzchj" podStartSLOduration=1.018591388 podStartE2EDuration="1.018591388s" podCreationTimestamp="2026-04-17 09:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:22:57.017492944 +0000 UTC m=+109.858530157" watchObservedRunningTime="2026-04-17 09:22:57.018591388 +0000 UTC m=+109.859628599" Apr 17 09:22:57.626870 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.626800 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:57.627069 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:57.627028 2569 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 09:22:57.627131 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:57.627096 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls podName:8f81e523-bc6a-40b3-8e85-891b04d3707b nodeName:}" failed. No retries permitted until 2026-04-17 09:22:59.627074693 +0000 UTC m=+112.468111898 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-pdqft" (UID: "8f81e523-bc6a-40b3-8e85-891b04d3707b") : secret "samples-operator-tls" not found Apr 17 09:22:57.831662 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.831618 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:57.831853 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:57.831779 2569 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:22:57.831853 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:57.831800 2569 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-75d68f6767-9txgs: secret "image-registry-tls" not found Apr 17 09:22:57.831969 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:57.831885 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls podName:414546fe-e0a7-4aad-bf7e-fa775be0036d nodeName:}" failed. No retries permitted until 2026-04-17 09:22:59.831865439 +0000 UTC m=+112.672902644 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls") pod "image-registry-75d68f6767-9txgs" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d") : secret "image-registry-tls" not found Apr 17 09:22:57.932330 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:57.932252 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:57.932489 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:57.932429 2569 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 09:22:57.932534 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:57.932511 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls podName:22a71e37-6733-4843-a7c4-2027bbd017e5 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:59.932489002 +0000 UTC m=+112.773526219 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-9rpfg" (UID: "22a71e37-6733-4843-a7c4-2027bbd017e5") : secret "cluster-monitoring-operator-tls" not found Apr 17 09:22:59.010065 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.010036 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/0.log" Apr 17 09:22:59.010498 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.010087 2569 generic.go:358] "Generic (PLEG): container finished" podID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" containerID="2b311672e212708cca545280bc0a1d8601dfb0e7e1d43f5e74292e593a61e201" exitCode=255 Apr 17 09:22:59.010498 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.010202 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" event={"ID":"ef4e2c76-47f7-42fb-a0e4-d95b609d0101","Type":"ContainerDied","Data":"2b311672e212708cca545280bc0a1d8601dfb0e7e1d43f5e74292e593a61e201"} Apr 17 09:22:59.010498 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.010398 2569 scope.go:117] "RemoveContainer" containerID="2b311672e212708cca545280bc0a1d8601dfb0e7e1d43f5e74292e593a61e201" Apr 17 09:22:59.011554 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.011533 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" event={"ID":"89378780-0d1b-40cd-930c-da0d27751523","Type":"ContainerStarted","Data":"8f18929d00b67b96c922cd27356f0c5055a0b46b88bd16adb289c9bf7a50816e"} Apr 17 09:22:59.038943 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.038898 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" podStartSLOduration=0.837187378 podStartE2EDuration="3.038884754s" podCreationTimestamp="2026-04-17 09:22:56 +0000 UTC" firstStartedPulling="2026-04-17 09:22:56.568115957 +0000 UTC m=+109.409153148" lastFinishedPulling="2026-04-17 09:22:58.769813335 +0000 UTC m=+111.610850524" observedRunningTime="2026-04-17 09:22:59.037987746 +0000 UTC m=+111.879024967" watchObservedRunningTime="2026-04-17 09:22:59.038884754 +0000 UTC m=+111.879921966" Apr 17 09:22:59.645001 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.644954 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:22:59.645178 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:59.645097 2569 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 09:22:59.645178 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:59.645160 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls podName:8f81e523-bc6a-40b3-8e85-891b04d3707b nodeName:}" failed. No retries permitted until 2026-04-17 09:23:03.645145984 +0000 UTC m=+116.486183173 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-pdqft" (UID: "8f81e523-bc6a-40b3-8e85-891b04d3707b") : secret "samples-operator-tls" not found Apr 17 09:22:59.846927 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.846894 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:22:59.847122 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:59.847065 2569 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:22:59.847122 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:59.847088 2569 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-75d68f6767-9txgs: secret "image-registry-tls" not found Apr 17 09:22:59.847244 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:59.847152 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls podName:414546fe-e0a7-4aad-bf7e-fa775be0036d nodeName:}" failed. No retries permitted until 2026-04-17 09:23:03.847132449 +0000 UTC m=+116.688169639 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls") pod "image-registry-75d68f6767-9txgs" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d") : secret "image-registry-tls" not found Apr 17 09:22:59.948322 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:22:59.948239 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:22:59.948454 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:59.948381 2569 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 09:22:59.948492 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:22:59.948455 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls podName:22a71e37-6733-4843-a7c4-2027bbd017e5 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:03.948431455 +0000 UTC m=+116.789468645 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-9rpfg" (UID: "22a71e37-6733-4843-a7c4-2027bbd017e5") : secret "cluster-monitoring-operator-tls" not found Apr 17 09:23:00.015309 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:00.015286 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/1.log" Apr 17 09:23:00.015674 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:00.015655 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/0.log" Apr 17 09:23:00.015722 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:00.015687 2569 generic.go:358] "Generic (PLEG): container finished" podID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" containerID="694823e9498b11cf105d5a19c106a6de0a332ac73938a326a42df4217ce8cea8" exitCode=255 Apr 17 09:23:00.015794 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:00.015770 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" event={"ID":"ef4e2c76-47f7-42fb-a0e4-d95b609d0101","Type":"ContainerDied","Data":"694823e9498b11cf105d5a19c106a6de0a332ac73938a326a42df4217ce8cea8"} Apr 17 09:23:00.015879 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:00.015831 2569 scope.go:117] "RemoveContainer" containerID="2b311672e212708cca545280bc0a1d8601dfb0e7e1d43f5e74292e593a61e201" Apr 17 09:23:00.016167 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:00.016149 2569 scope.go:117] "RemoveContainer" containerID="694823e9498b11cf105d5a19c106a6de0a332ac73938a326a42df4217ce8cea8" Apr 17 09:23:00.016326 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:00.016309 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-z9nkj_openshift-console-operator(ef4e2c76-47f7-42fb-a0e4-d95b609d0101)\"" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" podUID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" Apr 17 09:23:01.018511 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:01.018487 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/1.log" Apr 17 09:23:01.018904 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:01.018806 2569 scope.go:117] "RemoveContainer" containerID="694823e9498b11cf105d5a19c106a6de0a332ac73938a326a42df4217ce8cea8" Apr 17 09:23:01.019003 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:01.018977 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-z9nkj_openshift-console-operator(ef4e2c76-47f7-42fb-a0e4-d95b609d0101)\"" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" podUID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" Apr 17 09:23:03.406328 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:03.406300 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-b5dkm_9fcd330e-f463-420a-a91c-f2a6e23daebc/dns-node-resolver/0.log" Apr 17 09:23:03.679814 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:03.679730 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:23:03.679963 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:03.679855 2569 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 09:23:03.679963 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:03.679909 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls podName:8f81e523-bc6a-40b3-8e85-891b04d3707b nodeName:}" failed. No retries permitted until 2026-04-17 09:23:11.679895045 +0000 UTC m=+124.520932235 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-pdqft" (UID: "8f81e523-bc6a-40b3-8e85-891b04d3707b") : secret "samples-operator-tls" not found Apr 17 09:23:03.880588 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:03.880558 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:23:03.880727 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:03.880711 2569 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:23:03.880766 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:03.880731 2569 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-75d68f6767-9txgs: secret "image-registry-tls" not found Apr 17 09:23:03.880799 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:03.880788 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls podName:414546fe-e0a7-4aad-bf7e-fa775be0036d nodeName:}" failed. No retries permitted until 2026-04-17 09:23:11.88077105 +0000 UTC m=+124.721808242 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls") pod "image-registry-75d68f6767-9txgs" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d") : secret "image-registry-tls" not found Apr 17 09:23:03.981649 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:03.981578 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:23:03.981770 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:03.981725 2569 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 09:23:03.981808 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:03.981785 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls podName:22a71e37-6733-4843-a7c4-2027bbd017e5 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:11.981771374 +0000 UTC m=+124.822808564 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-9rpfg" (UID: "22a71e37-6733-4843-a7c4-2027bbd017e5") : secret "cluster-monitoring-operator-tls" not found Apr 17 09:23:04.605176 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:04.605145 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-t5ckx_09912815-97d7-47ef-939f-9f3815c992c1/node-ca/0.log" Apr 17 09:23:06.006868 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:06.006838 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-zj5rl_89378780-0d1b-40cd-930c-da0d27751523/kube-storage-version-migrator-operator/0.log" Apr 17 09:23:06.286933 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:06.286852 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:23:06.286933 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:06.286893 2569 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:23:06.287242 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:06.287230 2569 scope.go:117] "RemoveContainer" containerID="694823e9498b11cf105d5a19c106a6de0a332ac73938a326a42df4217ce8cea8" Apr 17 09:23:06.287405 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:06.287382 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-z9nkj_openshift-console-operator(ef4e2c76-47f7-42fb-a0e4-d95b609d0101)\"" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" podUID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" Apr 17 09:23:11.741213 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:11.741176 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:23:11.743494 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:11.743465 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f81e523-bc6a-40b3-8e85-891b04d3707b-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-pdqft\" (UID: \"8f81e523-bc6a-40b3-8e85-891b04d3707b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:23:11.881315 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:11.881284 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-6vwhv\"" Apr 17 09:23:11.889964 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:11.889941 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" Apr 17 09:23:11.942887 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:11.942847 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:23:11.945633 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:11.945604 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"image-registry-75d68f6767-9txgs\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:23:12.003195 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:12.003121 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft"] Apr 17 09:23:12.043914 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:12.043892 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:23:12.044027 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:12.043999 2569 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 09:23:12.044068 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:12.044044 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls podName:22a71e37-6733-4843-a7c4-2027bbd017e5 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:28.044029664 +0000 UTC m=+140.885066854 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-9rpfg" (UID: "22a71e37-6733-4843-a7c4-2027bbd017e5") : secret "cluster-monitoring-operator-tls" not found Apr 17 09:23:12.054330 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:12.054311 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-brmrj\"" Apr 17 09:23:12.063076 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:12.063058 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:23:12.171862 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:12.171746 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-75d68f6767-9txgs"] Apr 17 09:23:12.176529 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:12.176500 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod414546fe_e0a7_4aad_bf7e_fa775be0036d.slice/crio-f698f601d7d4f3813728e0633ef14583d5358fb8fe657b9ed916b6f2e79d82fd WatchSource:0}: Error finding container f698f601d7d4f3813728e0633ef14583d5358fb8fe657b9ed916b6f2e79d82fd: Status 404 returned error can't find the container with id f698f601d7d4f3813728e0633ef14583d5358fb8fe657b9ed916b6f2e79d82fd Apr 17 09:23:13.043993 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:13.043957 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" event={"ID":"414546fe-e0a7-4aad-bf7e-fa775be0036d","Type":"ContainerStarted","Data":"dbe56c5127783c76b76a8a2c1ffb0ee70ae10f29a71eb4b0ec02a3aff4bb06bc"} Apr 17 09:23:13.043993 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:13.044000 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" event={"ID":"414546fe-e0a7-4aad-bf7e-fa775be0036d","Type":"ContainerStarted","Data":"f698f601d7d4f3813728e0633ef14583d5358fb8fe657b9ed916b6f2e79d82fd"} Apr 17 09:23:13.044471 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:13.044073 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:23:13.044900 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:13.044874 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" event={"ID":"8f81e523-bc6a-40b3-8e85-891b04d3707b","Type":"ContainerStarted","Data":"44b29d645bc8b4f45b15efae6afd1044c3f6a0e1a1173b9da1cc77b8d6287002"} Apr 17 09:23:13.062860 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:13.062802 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" podStartSLOduration=17.062791706 podStartE2EDuration="17.062791706s" podCreationTimestamp="2026-04-17 09:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:23:13.061163211 +0000 UTC m=+125.902200422" watchObservedRunningTime="2026-04-17 09:23:13.062791706 +0000 UTC m=+125.903828918" Apr 17 09:23:15.051230 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:15.051186 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" event={"ID":"8f81e523-bc6a-40b3-8e85-891b04d3707b","Type":"ContainerStarted","Data":"38bfcff44f4784ec056fa0de9e2dea607fac9b6d102ac55acb9c586a42900f8c"} Apr 17 09:23:15.051230 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:15.051226 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" event={"ID":"8f81e523-bc6a-40b3-8e85-891b04d3707b","Type":"ContainerStarted","Data":"153c5e2cb66a7fbc6edf068312fa5ed0629f0ad060551670ff20da76ed2f250f"} Apr 17 09:23:15.066756 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:15.066710 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-pdqft" podStartSLOduration=18.094939157 podStartE2EDuration="20.06669746s" podCreationTimestamp="2026-04-17 09:22:55 +0000 UTC" firstStartedPulling="2026-04-17 09:23:12.042387602 +0000 UTC m=+124.883424792" lastFinishedPulling="2026-04-17 09:23:14.014145903 +0000 UTC m=+126.855183095" observedRunningTime="2026-04-17 09:23:15.066112448 +0000 UTC m=+127.907149659" watchObservedRunningTime="2026-04-17 09:23:15.06669746 +0000 UTC m=+127.907734705" Apr 17 09:23:17.386750 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:17.386716 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:23:17.389547 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:17.389520 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/833e0b85-0aba-4f60-bbf6-a215ed02831a-metrics-certs\") pod \"network-metrics-daemon-4kbg5\" (UID: \"833e0b85-0aba-4f60-bbf6-a215ed02831a\") " pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:23:17.522866 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:17.522838 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-8wjkc\"" Apr 17 09:23:17.530714 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:17.530684 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4kbg5" Apr 17 09:23:17.638975 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:17.638909 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4kbg5"] Apr 17 09:23:17.641682 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:17.641655 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod833e0b85_0aba_4f60_bbf6_a215ed02831a.slice/crio-6b59e75d8642e6a61922012281306fca60a86a4165dbf5d718a41c60a44e5fe3 WatchSource:0}: Error finding container 6b59e75d8642e6a61922012281306fca60a86a4165dbf5d718a41c60a44e5fe3: Status 404 returned error can't find the container with id 6b59e75d8642e6a61922012281306fca60a86a4165dbf5d718a41c60a44e5fe3 Apr 17 09:23:18.059653 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:18.059622 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4kbg5" event={"ID":"833e0b85-0aba-4f60-bbf6-a215ed02831a","Type":"ContainerStarted","Data":"6b59e75d8642e6a61922012281306fca60a86a4165dbf5d718a41c60a44e5fe3"} Apr 17 09:23:19.065749 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:19.065715 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4kbg5" event={"ID":"833e0b85-0aba-4f60-bbf6-a215ed02831a","Type":"ContainerStarted","Data":"4dcc54f9190726e6927edbd67e5a89f37841c9827a623d55ffe021ae99e7e40b"} Apr 17 09:23:19.065749 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:19.065751 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4kbg5" event={"ID":"833e0b85-0aba-4f60-bbf6-a215ed02831a","Type":"ContainerStarted","Data":"7226d94861fe5ad822c09f35804a25a014f9999794e38ed27831ce4e535c8b42"} Apr 17 09:23:19.083021 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:19.082076 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4kbg5" podStartSLOduration=131.069089249 podStartE2EDuration="2m12.08205953s" podCreationTimestamp="2026-04-17 09:21:07 +0000 UTC" firstStartedPulling="2026-04-17 09:23:17.643467542 +0000 UTC m=+130.484504731" lastFinishedPulling="2026-04-17 09:23:18.656437809 +0000 UTC m=+131.497475012" observedRunningTime="2026-04-17 09:23:19.08002486 +0000 UTC m=+131.921062073" watchObservedRunningTime="2026-04-17 09:23:19.08205953 +0000 UTC m=+131.923096744" Apr 17 09:23:20.702581 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:20.702547 2569 scope.go:117] "RemoveContainer" containerID="694823e9498b11cf105d5a19c106a6de0a332ac73938a326a42df4217ce8cea8" Apr 17 09:23:21.072460 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:21.072434 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/2.log" Apr 17 09:23:21.072812 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:21.072796 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/1.log" Apr 17 09:23:21.072883 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:21.072846 2569 generic.go:358] "Generic (PLEG): container finished" podID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" containerID="eba2cd46dafcad700c8f2567c593d0b9d672409d8151ffd24bc9bf55690d36cc" exitCode=255 Apr 17 09:23:21.072924 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:21.072912 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" event={"ID":"ef4e2c76-47f7-42fb-a0e4-d95b609d0101","Type":"ContainerDied","Data":"eba2cd46dafcad700c8f2567c593d0b9d672409d8151ffd24bc9bf55690d36cc"} Apr 17 09:23:21.072971 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:21.072943 2569 scope.go:117] "RemoveContainer" containerID="694823e9498b11cf105d5a19c106a6de0a332ac73938a326a42df4217ce8cea8" Apr 17 09:23:21.073227 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:21.073207 2569 scope.go:117] "RemoveContainer" containerID="eba2cd46dafcad700c8f2567c593d0b9d672409d8151ffd24bc9bf55690d36cc" Apr 17 09:23:21.073456 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:21.073437 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-z9nkj_openshift-console-operator(ef4e2c76-47f7-42fb-a0e4-d95b609d0101)\"" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" podUID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" Apr 17 09:23:22.076632 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:22.076604 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/2.log" Apr 17 09:23:25.066437 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.066393 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-75d68f6767-9txgs"] Apr 17 09:23:25.104280 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.104254 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-fcffd4c79-f7kd9"] Apr 17 09:23:25.108209 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.108183 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-rfxfc"] Apr 17 09:23:25.108362 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.108348 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.111212 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.111187 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.114192 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.114166 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 17 09:23:25.114302 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.114239 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 17 09:23:25.114302 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.114286 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 17 09:23:25.114452 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.114434 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-fqvbv\"" Apr 17 09:23:25.116121 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.116068 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 17 09:23:25.120766 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.120742 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-fcffd4c79-f7kd9"] Apr 17 09:23:25.121616 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.121592 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-rfxfc"] Apr 17 09:23:25.240756 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.240725 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-installation-pull-secrets\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.240951 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.240764 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.240951 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.240786 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nflkk\" (UniqueName: \"kubernetes.io/projected/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-kube-api-access-nflkk\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.240951 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.240889 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-registry-tls\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.240951 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.240934 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-image-registry-private-configuration\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.241096 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.240977 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-registry-certificates\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.241096 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.241009 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-bound-sa-token\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.241096 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.241024 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-ca-trust-extracted\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.241096 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.241056 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-data-volume\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.241096 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.241073 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.241096 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.241094 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-trusted-ca\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.241296 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.241117 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-crio-socket\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.241296 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.241149 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wlt\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-kube-api-access-67wlt\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.341776 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.341697 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-registry-certificates\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.341776 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.341743 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-bound-sa-token\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.341776 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.341761 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-ca-trust-extracted\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.341776 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.341779 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-data-volume\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.342090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.341913 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.342090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.341971 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-trusted-ca\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.342090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342002 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-crio-socket\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.342090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342047 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-67wlt\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-kube-api-access-67wlt\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.342090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342079 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-installation-pull-secrets\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.342327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342106 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.342327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342128 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-data-volume\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.342327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342142 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nflkk\" (UniqueName: \"kubernetes.io/projected/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-kube-api-access-nflkk\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.342327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342205 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-registry-tls\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.342327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342226 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-ca-trust-extracted\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.342327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342263 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-image-registry-private-configuration\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.342610 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342378 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-crio-socket\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.342610 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.342523 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.343095 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.343065 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-registry-certificates\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.343392 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.343367 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-trusted-ca\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.344734 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.344708 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-installation-pull-secrets\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.344863 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.344757 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.344964 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.344945 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-registry-tls\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.344964 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.344955 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-image-registry-private-configuration\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.350127 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.350105 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-bound-sa-token\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.350451 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.350422 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wlt\" (UniqueName: \"kubernetes.io/projected/5ab3de37-fccc-4ce9-9c52-a0d690b6311a-kube-api-access-67wlt\") pod \"image-registry-fcffd4c79-f7kd9\" (UID: \"5ab3de37-fccc-4ce9-9c52-a0d690b6311a\") " pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.350764 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.350740 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nflkk\" (UniqueName: \"kubernetes.io/projected/3d38afde-3d74-46a3-a5f7-e4cec8ee56d6-kube-api-access-nflkk\") pod \"insights-runtime-extractor-rfxfc\" (UID: \"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6\") " pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.423946 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.423918 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:25.427671 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.427653 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-rfxfc" Apr 17 09:23:25.552518 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.552487 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-fcffd4c79-f7kd9"] Apr 17 09:23:25.555184 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:25.555155 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ab3de37_fccc_4ce9_9c52_a0d690b6311a.slice/crio-f41a006287a7c588be1546dfa7e976c3b87c9962312347ddde638eca5cecb298 WatchSource:0}: Error finding container f41a006287a7c588be1546dfa7e976c3b87c9962312347ddde638eca5cecb298: Status 404 returned error can't find the container with id f41a006287a7c588be1546dfa7e976c3b87c9962312347ddde638eca5cecb298 Apr 17 09:23:25.564167 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:25.564140 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-rfxfc"] Apr 17 09:23:25.567111 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:25.567084 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d38afde_3d74_46a3_a5f7_e4cec8ee56d6.slice/crio-71433301b0bf1e8ac19dd6ce179939de7bd645e882318e471a45db18d06a6e4d WatchSource:0}: Error finding container 71433301b0bf1e8ac19dd6ce179939de7bd645e882318e471a45db18d06a6e4d: Status 404 returned error can't find the container with id 71433301b0bf1e8ac19dd6ce179939de7bd645e882318e471a45db18d06a6e4d Apr 17 09:23:26.087859 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.087805 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rfxfc" event={"ID":"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6","Type":"ContainerStarted","Data":"c371f86311261d6aea9740660a948c868d7e43baf66014f20bc6e7eed8afe236"} Apr 17 09:23:26.088246 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.087868 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rfxfc" event={"ID":"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6","Type":"ContainerStarted","Data":"71433301b0bf1e8ac19dd6ce179939de7bd645e882318e471a45db18d06a6e4d"} Apr 17 09:23:26.089042 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.089016 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" event={"ID":"5ab3de37-fccc-4ce9-9c52-a0d690b6311a","Type":"ContainerStarted","Data":"87ed3981a8f67dc136da3ea1e0619393571bfc499ac1dbd1d981d7bcb0277cc5"} Apr 17 09:23:26.089139 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.089049 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" event={"ID":"5ab3de37-fccc-4ce9-9c52-a0d690b6311a","Type":"ContainerStarted","Data":"f41a006287a7c588be1546dfa7e976c3b87c9962312347ddde638eca5cecb298"} Apr 17 09:23:26.089182 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.089142 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:26.108022 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.107978 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" podStartSLOduration=1.107962928 podStartE2EDuration="1.107962928s" podCreationTimestamp="2026-04-17 09:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:23:26.107036571 +0000 UTC m=+138.948073779" watchObservedRunningTime="2026-04-17 09:23:26.107962928 +0000 UTC m=+138.949000168" Apr 17 09:23:26.286918 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.286881 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:23:26.286918 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.286920 2569 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:23:26.287266 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:26.287249 2569 scope.go:117] "RemoveContainer" containerID="eba2cd46dafcad700c8f2567c593d0b9d672409d8151ffd24bc9bf55690d36cc" Apr 17 09:23:26.287439 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:26.287423 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-z9nkj_openshift-console-operator(ef4e2c76-47f7-42fb-a0e4-d95b609d0101)\"" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" podUID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" Apr 17 09:23:27.094441 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:27.094398 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rfxfc" event={"ID":"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6","Type":"ContainerStarted","Data":"b37a500a0ddbb558fdbeb5e3bccb7a1651a1230054148c58b2690ed6604972f1"} Apr 17 09:23:28.064830 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:28.064784 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:23:28.067059 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:28.067031 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a71e37-6733-4843-a7c4-2027bbd017e5-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-9rpfg\" (UID: \"22a71e37-6733-4843-a7c4-2027bbd017e5\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:23:28.098325 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:28.098299 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rfxfc" event={"ID":"3d38afde-3d74-46a3-a5f7-e4cec8ee56d6","Type":"ContainerStarted","Data":"1d6c52af1d8b305dc099ac7d36a371afc2c5c0053b112b80f625cba3febbe0f7"} Apr 17 09:23:28.114829 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:28.114775 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-rfxfc" podStartSLOduration=1.173305427 podStartE2EDuration="3.114760488s" podCreationTimestamp="2026-04-17 09:23:25 +0000 UTC" firstStartedPulling="2026-04-17 09:23:25.612964984 +0000 UTC m=+138.454002190" lastFinishedPulling="2026-04-17 09:23:27.55442006 +0000 UTC m=+140.395457251" observedRunningTime="2026-04-17 09:23:28.113280304 +0000 UTC m=+140.954317515" watchObservedRunningTime="2026-04-17 09:23:28.114760488 +0000 UTC m=+140.955797700" Apr 17 09:23:28.329359 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:28.329277 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-4xs4t\"" Apr 17 09:23:28.338081 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:28.338060 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" Apr 17 09:23:28.454470 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:28.454094 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg"] Apr 17 09:23:28.457521 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:28.457488 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22a71e37_6733_4843_a7c4_2027bbd017e5.slice/crio-2b9a50e535be7ce48f2dfb35ac09ecb4d34469161f8f552ea6d28859ce7e5ffa WatchSource:0}: Error finding container 2b9a50e535be7ce48f2dfb35ac09ecb4d34469161f8f552ea6d28859ce7e5ffa: Status 404 returned error can't find the container with id 2b9a50e535be7ce48f2dfb35ac09ecb4d34469161f8f552ea6d28859ce7e5ffa Apr 17 09:23:29.102634 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:29.102591 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" event={"ID":"22a71e37-6733-4843-a7c4-2027bbd017e5","Type":"ContainerStarted","Data":"2b9a50e535be7ce48f2dfb35ac09ecb4d34469161f8f552ea6d28859ce7e5ffa"} Apr 17 09:23:30.106195 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.106160 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" event={"ID":"22a71e37-6733-4843-a7c4-2027bbd017e5","Type":"ContainerStarted","Data":"71b50f12d004ae4d40b3af661cbc59b5757490076ea9559770af7c45a1a16712"} Apr 17 09:23:30.122519 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.122458 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-9rpfg" podStartSLOduration=32.584142337 podStartE2EDuration="34.12241254s" podCreationTimestamp="2026-04-17 09:22:56 +0000 UTC" firstStartedPulling="2026-04-17 09:23:28.459271607 +0000 UTC m=+141.300308796" lastFinishedPulling="2026-04-17 09:23:29.997541805 +0000 UTC m=+142.838578999" observedRunningTime="2026-04-17 09:23:30.121963734 +0000 UTC m=+142.963000959" watchObservedRunningTime="2026-04-17 09:23:30.12241254 +0000 UTC m=+142.963449753" Apr 17 09:23:30.463436 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.463403 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw"] Apr 17 09:23:30.466805 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.466785 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:30.469010 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.468988 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-tls\"" Apr 17 09:23:30.469099 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.468997 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-dockercfg-hxxlm\"" Apr 17 09:23:30.472454 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.472431 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw"] Apr 17 09:23:30.584906 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.584875 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-7gvjw\" (UID: \"0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:30.686189 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:30.686159 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-7gvjw\" (UID: \"0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:30.686325 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:30.686296 2569 secret.go:189] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Apr 17 09:23:30.686368 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:30.686355 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates podName:0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d nodeName:}" failed. No retries permitted until 2026-04-17 09:23:31.186341217 +0000 UTC m=+144.027378410 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates") pod "prometheus-operator-admission-webhook-57cf98b594-7gvjw" (UID: "0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d") : secret "prometheus-operator-admission-webhook-tls" not found Apr 17 09:23:31.189032 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:31.189003 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-7gvjw\" (UID: \"0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:31.189373 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:31.189107 2569 secret.go:189] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Apr 17 09:23:31.189373 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:31.189154 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates podName:0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d nodeName:}" failed. No retries permitted until 2026-04-17 09:23:32.189142075 +0000 UTC m=+145.030179265 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates") pod "prometheus-operator-admission-webhook-57cf98b594-7gvjw" (UID: "0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d") : secret "prometheus-operator-admission-webhook-tls" not found Apr 17 09:23:32.196216 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:32.196176 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-7gvjw\" (UID: \"0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:32.198566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:32.198543 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-7gvjw\" (UID: \"0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:32.276176 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:32.276153 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:32.396902 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:32.396861 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw"] Apr 17 09:23:32.399481 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:32.399456 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a8a1e02_ea5f_4784_9eaf_96a0ccf9068d.slice/crio-60973ff3f8b0bf592fcdd62c2c7ceddba01c3562f704c83b02fff12e5fc2c8de WatchSource:0}: Error finding container 60973ff3f8b0bf592fcdd62c2c7ceddba01c3562f704c83b02fff12e5fc2c8de: Status 404 returned error can't find the container with id 60973ff3f8b0bf592fcdd62c2c7ceddba01c3562f704c83b02fff12e5fc2c8de Apr 17 09:23:33.114026 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:33.113973 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" event={"ID":"0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d","Type":"ContainerStarted","Data":"60973ff3f8b0bf592fcdd62c2c7ceddba01c3562f704c83b02fff12e5fc2c8de"} Apr 17 09:23:34.117792 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:34.117758 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" event={"ID":"0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d","Type":"ContainerStarted","Data":"d3510843fa528897f62024b7b98b066c699e1726777dbff02413908a9e59c492"} Apr 17 09:23:34.118224 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:34.117993 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:34.122755 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:34.122730 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" Apr 17 09:23:34.132826 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:34.132781 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-7gvjw" podStartSLOduration=2.932083777 podStartE2EDuration="4.13276786s" podCreationTimestamp="2026-04-17 09:23:30 +0000 UTC" firstStartedPulling="2026-04-17 09:23:32.401249671 +0000 UTC m=+145.242286860" lastFinishedPulling="2026-04-17 09:23:33.601933744 +0000 UTC m=+146.442970943" observedRunningTime="2026-04-17 09:23:34.131586467 +0000 UTC m=+146.972623679" watchObservedRunningTime="2026-04-17 09:23:34.13276786 +0000 UTC m=+146.973805069" Apr 17 09:23:35.071919 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:35.071875 2569 patch_prober.go:28] interesting pod/image-registry-75d68f6767-9txgs container/registry namespace/openshift-image-registry: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 17 09:23:35.072070 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:35.071935 2569 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" podUID="414546fe-e0a7-4aad-bf7e-fa775be0036d" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 17 09:23:38.856315 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.856284 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq"] Apr 17 09:23:38.859566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.859548 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:38.862664 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.862641 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-tls\"" Apr 17 09:23:38.862664 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.862654 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-dockercfg-d2zlj\"" Apr 17 09:23:38.862833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.862646 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 17 09:23:38.862833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.862692 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-kube-rbac-proxy-config\"" Apr 17 09:23:38.872460 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.872441 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq"] Apr 17 09:23:38.904137 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.904111 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-fm9gq"] Apr 17 09:23:38.907256 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.907237 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.909510 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.909484 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 17 09:23:38.909510 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.909505 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-hwqkh\"" Apr 17 09:23:38.909764 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.909727 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 17 09:23:38.909862 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.909769 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 17 09:23:38.948742 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948719 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-root\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.948871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948748 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54961548-4e6d-49bc-8671-8265a524aba5-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:38.948871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948766 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4222015d-39a6-4f11-b504-7a3dfc2576cf-metrics-client-ca\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.948871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948790 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-sys\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.948871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948847 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrp95\" (UniqueName: \"kubernetes.io/projected/54961548-4e6d-49bc-8671-8265a524aba5-kube-api-access-rrp95\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:38.949033 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948902 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-textfile\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.949033 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948924 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.949033 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948977 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kccc5\" (UniqueName: \"kubernetes.io/projected/4222015d-39a6-4f11-b504-7a3dfc2576cf-kube-api-access-kccc5\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.949033 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.948998 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:38.949185 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.949029 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:38.949185 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.949078 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-accelerators-collector-config\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.949185 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.949104 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-wtmp\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:38.949185 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:38.949130 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-tls\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050206 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050179 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-tls\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050222 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-root\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050240 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54961548-4e6d-49bc-8671-8265a524aba5-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.050308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050262 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4222015d-39a6-4f11-b504-7a3dfc2576cf-metrics-client-ca\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050287 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-sys\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050290 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-root\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050546 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050313 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rrp95\" (UniqueName: \"kubernetes.io/projected/54961548-4e6d-49bc-8671-8265a524aba5-kube-api-access-rrp95\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.050546 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050347 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-textfile\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050546 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050375 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050546 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050406 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-sys\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050546 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050432 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kccc5\" (UniqueName: \"kubernetes.io/projected/4222015d-39a6-4f11-b504-7a3dfc2576cf-kube-api-access-kccc5\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050790 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050570 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.050790 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050611 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.050790 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050650 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-accelerators-collector-config\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050790 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:39.050667 2569 secret.go:189] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Apr 17 09:23:39.050790 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050688 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-wtmp\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050790 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050697 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-textfile\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.050790 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:39.050718 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-tls podName:54961548-4e6d-49bc-8671-8265a524aba5 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:39.550704743 +0000 UTC m=+152.391741933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-tls") pod "openshift-state-metrics-9d44df66c-bttbq" (UID: "54961548-4e6d-49bc-8671-8265a524aba5") : secret "openshift-state-metrics-tls" not found Apr 17 09:23:39.051130 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050848 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-wtmp\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.051130 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.050881 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4222015d-39a6-4f11-b504-7a3dfc2576cf-metrics-client-ca\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.051130 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.051036 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54961548-4e6d-49bc-8671-8265a524aba5-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.051223 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.051186 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-accelerators-collector-config\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.053116 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.053089 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.053259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.053238 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.053315 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.053301 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/4222015d-39a6-4f11-b504-7a3dfc2576cf-node-exporter-tls\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.059500 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.059478 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kccc5\" (UniqueName: \"kubernetes.io/projected/4222015d-39a6-4f11-b504-7a3dfc2576cf-kube-api-access-kccc5\") pod \"node-exporter-fm9gq\" (UID: \"4222015d-39a6-4f11-b504-7a3dfc2576cf\") " pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.059691 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.059677 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrp95\" (UniqueName: \"kubernetes.io/projected/54961548-4e6d-49bc-8671-8265a524aba5-kube-api-access-rrp95\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.215520 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.215444 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-fm9gq" Apr 17 09:23:39.223411 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:39.223381 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4222015d_39a6_4f11_b504_7a3dfc2576cf.slice/crio-c1e0ec847bbdf38d6faad8c94afce1a0141e9e8234d403ee294db05aec3d6f03 WatchSource:0}: Error finding container c1e0ec847bbdf38d6faad8c94afce1a0141e9e8234d403ee294db05aec3d6f03: Status 404 returned error can't find the container with id c1e0ec847bbdf38d6faad8c94afce1a0141e9e8234d403ee294db05aec3d6f03 Apr 17 09:23:39.555374 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.555344 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.557670 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.557648 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54961548-4e6d-49bc-8671-8265a524aba5-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-bttbq\" (UID: \"54961548-4e6d-49bc-8671-8265a524aba5\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.703169 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.703140 2569 scope.go:117] "RemoveContainer" containerID="eba2cd46dafcad700c8f2567c593d0b9d672409d8151ffd24bc9bf55690d36cc" Apr 17 09:23:39.703413 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:39.703383 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-z9nkj_openshift-console-operator(ef4e2c76-47f7-42fb-a0e4-d95b609d0101)\"" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" podUID="ef4e2c76-47f7-42fb-a0e4-d95b609d0101" Apr 17 09:23:39.768352 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.768324 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" Apr 17 09:23:39.901010 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.900972 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq"] Apr 17 09:23:39.936955 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.936899 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:23:39.941859 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.941809 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.944528 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.944456 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Apr 17 09:23:39.944528 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.944456 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Apr 17 09:23:39.944528 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.944462 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Apr 17 09:23:39.944528 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.944526 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-g6hlq\"" Apr 17 09:23:39.944833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.944462 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Apr 17 09:23:39.944833 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.944506 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Apr 17 09:23:39.945143 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.944964 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Apr 17 09:23:39.945143 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.945028 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Apr 17 09:23:39.945143 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.945048 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Apr 17 09:23:39.945143 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.945100 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Apr 17 09:23:39.956488 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.956455 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:23:39.959799 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.959775 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.959924 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.959841 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8wbx\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-kube-api-access-n8wbx\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.959924 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.959914 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960041 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.959953 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960041 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.959986 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-tls-assets\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960041 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960011 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-web-config\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960194 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960086 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960194 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960117 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960194 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960142 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960352 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960207 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-config-out\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960352 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960237 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960352 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960262 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:39.960352 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:39.960296 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-config-volume\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.016122 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:40.016082 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54961548_4e6d_49bc_8671_8265a524aba5.slice/crio-8d42bcc97f512a6f46e4c22d71ab4a27862f049652308745451a02d3a2d0d456 WatchSource:0}: Error finding container 8d42bcc97f512a6f46e4c22d71ab4a27862f049652308745451a02d3a2d0d456: Status 404 returned error can't find the container with id 8d42bcc97f512a6f46e4c22d71ab4a27862f049652308745451a02d3a2d0d456 Apr 17 09:23:40.060948 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.060919 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061076 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.060980 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061076 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061011 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061076 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061058 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-config-out\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061251 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061082 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061251 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061107 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061251 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061140 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-config-volume\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061251 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061183 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061251 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061216 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n8wbx\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-kube-api-access-n8wbx\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061251 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061244 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061568 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061275 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061568 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061303 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-tls-assets\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061568 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061335 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-web-config\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061568 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.061468 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.061782 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:40.061599 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle podName:7de2893b-b16a-4458-b369-a1832bca1744 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:40.56157922 +0000 UTC m=+153.402616412 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "alertmanager-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle") pod "alertmanager-main-0" (UID: "7de2893b-b16a-4458-b369-a1832bca1744") : configmap references non-existent config key: ca-bundle.crt Apr 17 09:23:40.062361 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.062234 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.062904 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:40.062691 2569 secret.go:189] Couldn't get secret openshift-monitoring/alertmanager-main-tls: secret "alertmanager-main-tls" not found Apr 17 09:23:40.062904 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:40.062752 2569 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls podName:7de2893b-b16a-4458-b369-a1832bca1744 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:40.562736304 +0000 UTC m=+153.403773502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-alertmanager-main-tls" (UniqueName: "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls") pod "alertmanager-main-0" (UID: "7de2893b-b16a-4458-b369-a1832bca1744") : secret "alertmanager-main-tls" not found Apr 17 09:23:40.064335 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.064309 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-config-volume\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.064432 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.064309 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-web-config\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.065313 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.065289 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.065716 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.065679 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.065837 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.065799 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.067433 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.067393 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-config-out\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.067529 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.067483 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-tls-assets\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.067529 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.067515 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.072495 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.072473 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8wbx\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-kube-api-access-n8wbx\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.134500 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.134470 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-fm9gq" event={"ID":"4222015d-39a6-4f11-b504-7a3dfc2576cf","Type":"ContainerStarted","Data":"c1e0ec847bbdf38d6faad8c94afce1a0141e9e8234d403ee294db05aec3d6f03"} Apr 17 09:23:40.136091 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.136067 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" event={"ID":"54961548-4e6d-49bc-8671-8265a524aba5","Type":"ContainerStarted","Data":"5b672adc246509d87d5a8021282f6bc086e9e71cf9f625a9a96c894089786d05"} Apr 17 09:23:40.136185 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.136100 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" event={"ID":"54961548-4e6d-49bc-8671-8265a524aba5","Type":"ContainerStarted","Data":"8d42bcc97f512a6f46e4c22d71ab4a27862f049652308745451a02d3a2d0d456"} Apr 17 09:23:40.565102 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.565072 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.565266 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.565135 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.566090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.566061 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.567486 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.567460 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.853458 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.853372 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:23:40.992971 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:40.992945 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:23:40.994723 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:40.994684 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7de2893b_b16a_4458_b369_a1832bca1744.slice/crio-98f9ce50fbaa76f81a6f7d78426a3223607f17ce40d8a5edc12f44c4196bf722 WatchSource:0}: Error finding container 98f9ce50fbaa76f81a6f7d78426a3223607f17ce40d8a5edc12f44c4196bf722: Status 404 returned error can't find the container with id 98f9ce50fbaa76f81a6f7d78426a3223607f17ce40d8a5edc12f44c4196bf722 Apr 17 09:23:41.140737 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.140644 2569 generic.go:358] "Generic (PLEG): container finished" podID="4222015d-39a6-4f11-b504-7a3dfc2576cf" containerID="3437eb8410ec4d3b609d9a68e5af76a6311d6979afe9c47e987e8830c3a3e8f5" exitCode=0 Apr 17 09:23:41.140932 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.140743 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-fm9gq" event={"ID":"4222015d-39a6-4f11-b504-7a3dfc2576cf","Type":"ContainerDied","Data":"3437eb8410ec4d3b609d9a68e5af76a6311d6979afe9c47e987e8830c3a3e8f5"} Apr 17 09:23:41.141948 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.141920 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"98f9ce50fbaa76f81a6f7d78426a3223607f17ce40d8a5edc12f44c4196bf722"} Apr 17 09:23:41.143626 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.143601 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" event={"ID":"54961548-4e6d-49bc-8671-8265a524aba5","Type":"ContainerStarted","Data":"0e292e42762d3b2e0a05e9ed4a2cfdb47f8fcb2a47000a67407f7dabee0d86bb"} Apr 17 09:23:41.864434 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.864400 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-577f9447f9-8p9fd"] Apr 17 09:23:41.868537 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.868517 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.871057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.871019 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-web\"" Apr 17 09:23:41.871174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.871089 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-metrics\"" Apr 17 09:23:41.871174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.871135 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-grpc-tls-6o6iu98tv380n\"" Apr 17 09:23:41.871296 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.871189 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-rules\"" Apr 17 09:23:41.871386 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.871353 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-tls\"" Apr 17 09:23:41.871501 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.871429 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-dockercfg-z7x65\"" Apr 17 09:23:41.871501 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.871485 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy\"" Apr 17 09:23:41.879930 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.879892 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-577f9447f9-8p9fd"] Apr 17 09:23:41.979981 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.979948 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.979981 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.979984 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/52442855-76a8-4ca3-a454-53fe40b0e7d0-metrics-client-ca\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.980214 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.980037 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-grpc-tls\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.980214 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.980091 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.980214 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.980166 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjp8b\" (UniqueName: \"kubernetes.io/projected/52442855-76a8-4ca3-a454-53fe40b0e7d0-kube-api-access-vjp8b\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.980375 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.980255 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.980375 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.980276 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-tls\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:41.980375 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:41.980297 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.080715 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.080688 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.081057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.080732 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vjp8b\" (UniqueName: \"kubernetes.io/projected/52442855-76a8-4ca3-a454-53fe40b0e7d0-kube-api-access-vjp8b\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.081057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.080831 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.081057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.080873 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-tls\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.081057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.080904 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.081057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.080942 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.081057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.080979 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/52442855-76a8-4ca3-a454-53fe40b0e7d0-metrics-client-ca\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.081057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.081033 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-grpc-tls\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.082222 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.081751 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/52442855-76a8-4ca3-a454-53fe40b0e7d0-metrics-client-ca\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.083530 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.083499 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.083634 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.083582 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-tls\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.083698 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.083639 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.084193 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.084168 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-grpc-tls\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.084285 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.084215 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.084483 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.084464 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/52442855-76a8-4ca3-a454-53fe40b0e7d0-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.087565 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.087541 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjp8b\" (UniqueName: \"kubernetes.io/projected/52442855-76a8-4ca3-a454-53fe40b0e7d0-kube-api-access-vjp8b\") pod \"thanos-querier-577f9447f9-8p9fd\" (UID: \"52442855-76a8-4ca3-a454-53fe40b0e7d0\") " pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.148093 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.148063 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-fm9gq" event={"ID":"4222015d-39a6-4f11-b504-7a3dfc2576cf","Type":"ContainerStarted","Data":"c4ee31b0b604befd9b0c1a99a068a058ea2cea2736fd6e21132330c756e05501"} Apr 17 09:23:42.148201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.148103 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-fm9gq" event={"ID":"4222015d-39a6-4f11-b504-7a3dfc2576cf","Type":"ContainerStarted","Data":"a58056f8abfc7a80928c7cb3b8f7660ecbf8e95f96d1560988bc5d75120fc68f"} Apr 17 09:23:42.149535 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.149505 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310"} Apr 17 09:23:42.151376 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.151354 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" event={"ID":"54961548-4e6d-49bc-8671-8265a524aba5","Type":"ContainerStarted","Data":"75281283658d173a18508fff3b3bd2380bd23c2b3551e1dce8af806335c16e3a"} Apr 17 09:23:42.168970 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.168931 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-fm9gq" podStartSLOduration=3.332276878 podStartE2EDuration="4.168917532s" podCreationTimestamp="2026-04-17 09:23:38 +0000 UTC" firstStartedPulling="2026-04-17 09:23:39.225110753 +0000 UTC m=+152.066147948" lastFinishedPulling="2026-04-17 09:23:40.061751407 +0000 UTC m=+152.902788602" observedRunningTime="2026-04-17 09:23:42.167445478 +0000 UTC m=+155.008482690" watchObservedRunningTime="2026-04-17 09:23:42.168917532 +0000 UTC m=+155.009954744" Apr 17 09:23:42.181871 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.181849 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:42.183559 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.183484 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-bttbq" podStartSLOduration=3.040806844 podStartE2EDuration="4.18346771s" podCreationTimestamp="2026-04-17 09:23:38 +0000 UTC" firstStartedPulling="2026-04-17 09:23:40.14079268 +0000 UTC m=+152.981829870" lastFinishedPulling="2026-04-17 09:23:41.283453529 +0000 UTC m=+154.124490736" observedRunningTime="2026-04-17 09:23:42.183461157 +0000 UTC m=+155.024498371" watchObservedRunningTime="2026-04-17 09:23:42.18346771 +0000 UTC m=+155.024504923" Apr 17 09:23:42.296153 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:42.296126 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-577f9447f9-8p9fd"] Apr 17 09:23:42.299067 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:42.299032 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52442855_76a8_4ca3_a454_53fe40b0e7d0.slice/crio-e04399e48354b54885dc2e0da8f589d55d05e3a53d6014a826aa4b5bbb1a1539 WatchSource:0}: Error finding container e04399e48354b54885dc2e0da8f589d55d05e3a53d6014a826aa4b5bbb1a1539: Status 404 returned error can't find the container with id e04399e48354b54885dc2e0da8f589d55d05e3a53d6014a826aa4b5bbb1a1539 Apr 17 09:23:43.157380 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.157335 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" event={"ID":"52442855-76a8-4ca3-a454-53fe40b0e7d0","Type":"ContainerStarted","Data":"e04399e48354b54885dc2e0da8f589d55d05e3a53d6014a826aa4b5bbb1a1539"} Apr 17 09:23:43.159879 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.159846 2569 generic.go:358] "Generic (PLEG): container finished" podID="7de2893b-b16a-4458-b369-a1832bca1744" containerID="8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310" exitCode=0 Apr 17 09:23:43.160123 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.159922 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310"} Apr 17 09:23:43.180584 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.180558 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-544766d54-ktmxf"] Apr 17 09:23:43.183873 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.183854 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.186627 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.186335 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-server-audit-profiles\"" Apr 17 09:23:43.186799 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.186775 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kubelet-serving-ca-bundle\"" Apr 17 09:23:43.186963 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.186919 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-tls\"" Apr 17 09:23:43.187059 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.187037 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-2h7onim92n8qg\"" Apr 17 09:23:43.187280 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.187263 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-client-certs\"" Apr 17 09:23:43.187503 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.187474 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-dockercfg-4xz57\"" Apr 17 09:23:43.191438 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.191418 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-544766d54-ktmxf"] Apr 17 09:23:43.290656 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.290620 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-secret-metrics-server-tls\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.290809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.290662 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a45e7480-74e1-407a-85ac-03d53fb98a59-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.290809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.290693 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-client-ca-bundle\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.290809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.290723 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a45e7480-74e1-407a-85ac-03d53fb98a59-metrics-server-audit-profiles\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.290983 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.290857 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b9mh\" (UniqueName: \"kubernetes.io/projected/a45e7480-74e1-407a-85ac-03d53fb98a59-kube-api-access-9b9mh\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.290983 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.290927 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a45e7480-74e1-407a-85ac-03d53fb98a59-audit-log\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.291063 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.290993 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-secret-metrics-server-client-certs\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.391679 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.391644 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a45e7480-74e1-407a-85ac-03d53fb98a59-audit-log\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.391856 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.391712 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-secret-metrics-server-client-certs\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.391856 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.391772 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-secret-metrics-server-tls\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.391856 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.391800 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a45e7480-74e1-407a-85ac-03d53fb98a59-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.391856 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.391849 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-client-ca-bundle\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.392059 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.391879 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a45e7480-74e1-407a-85ac-03d53fb98a59-metrics-server-audit-profiles\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.392059 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.392007 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9b9mh\" (UniqueName: \"kubernetes.io/projected/a45e7480-74e1-407a-85ac-03d53fb98a59-kube-api-access-9b9mh\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.392158 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.392102 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a45e7480-74e1-407a-85ac-03d53fb98a59-audit-log\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.392998 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.392953 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a45e7480-74e1-407a-85ac-03d53fb98a59-metrics-server-audit-profiles\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.393153 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.393133 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a45e7480-74e1-407a-85ac-03d53fb98a59-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.394693 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.394670 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-secret-metrics-server-client-certs\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.394862 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.394842 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-client-ca-bundle\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.394948 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.394928 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a45e7480-74e1-407a-85ac-03d53fb98a59-secret-metrics-server-tls\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.400089 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.400066 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b9mh\" (UniqueName: \"kubernetes.io/projected/a45e7480-74e1-407a-85ac-03d53fb98a59-kube-api-access-9b9mh\") pod \"metrics-server-544766d54-ktmxf\" (UID: \"a45e7480-74e1-407a-85ac-03d53fb98a59\") " pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.497675 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.497608 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:23:43.621190 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:43.621157 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-544766d54-ktmxf"] Apr 17 09:23:44.042539 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:44.042493 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-27fm4" podUID="b171f472-b404-4138-bf91-727a857e3224" Apr 17 09:23:44.052774 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:23:44.052740 2569 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-pf795" podUID="e5c6f5f4-e252-4726-9992-25c01cae77a3" Apr 17 09:23:44.163488 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:44.163458 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-27fm4" Apr 17 09:23:44.204550 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:44.204524 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda45e7480_74e1_407a_85ac_03d53fb98a59.slice/crio-ce0924726f05bb7656096b0439549a1047282a34e9d35722844faf0b5104d82a WatchSource:0}: Error finding container ce0924726f05bb7656096b0439549a1047282a34e9d35722844faf0b5104d82a: Status 404 returned error can't find the container with id ce0924726f05bb7656096b0439549a1047282a34e9d35722844faf0b5104d82a Apr 17 09:23:45.071045 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.070999 2569 patch_prober.go:28] interesting pod/image-registry-75d68f6767-9txgs container/registry namespace/openshift-image-registry: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 17 09:23:45.071229 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.071077 2569 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" podUID="414546fe-e0a7-4aad-bf7e-fa775be0036d" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 17 09:23:45.169726 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.169641 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30"} Apr 17 09:23:45.169726 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.169683 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5"} Apr 17 09:23:45.169726 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.169700 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7"} Apr 17 09:23:45.169726 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.169714 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683"} Apr 17 09:23:45.169726 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.169725 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9"} Apr 17 09:23:45.171037 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.170995 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" event={"ID":"a45e7480-74e1-407a-85ac-03d53fb98a59","Type":"ContainerStarted","Data":"ce0924726f05bb7656096b0439549a1047282a34e9d35722844faf0b5104d82a"} Apr 17 09:23:45.172946 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.172918 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" event={"ID":"52442855-76a8-4ca3-a454-53fe40b0e7d0","Type":"ContainerStarted","Data":"f6b099ebee93e30333d314ccb83ec538dce37ccac0e3fc53a92b65caa8b968cd"} Apr 17 09:23:45.173043 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.172951 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" event={"ID":"52442855-76a8-4ca3-a454-53fe40b0e7d0","Type":"ContainerStarted","Data":"2260aeb56ed36ba2c7583ea93163f87ded530c1540d148e6c07c6dc41797986c"} Apr 17 09:23:45.173043 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.172967 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" event={"ID":"52442855-76a8-4ca3-a454-53fe40b0e7d0","Type":"ContainerStarted","Data":"71a5e7fd822a552ca7937b807150fa47deb683056a02ece760d506bc76c7ab6c"} Apr 17 09:23:45.430383 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.430267 2569 patch_prober.go:28] interesting pod/image-registry-fcffd4c79-f7kd9 container/registry namespace/openshift-image-registry: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 17 09:23:45.430383 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:45.430327 2569 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" podUID="5ab3de37-fccc-4ce9-9c52-a0d690b6311a" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 17 09:23:46.178020 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:46.177990 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" event={"ID":"52442855-76a8-4ca3-a454-53fe40b0e7d0","Type":"ContainerStarted","Data":"769e049b7fce13cb1d8deda4a7c183b1d7cf6344d7adfa4975b1b74dc160248f"} Apr 17 09:23:46.178336 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:46.178033 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" event={"ID":"52442855-76a8-4ca3-a454-53fe40b0e7d0","Type":"ContainerStarted","Data":"5bdb401d8c3d7714a792ff3afe801666ee05b48fd009f5cb657e60966accd1a8"} Apr 17 09:23:46.180906 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:46.180804 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerStarted","Data":"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e"} Apr 17 09:23:46.182044 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:46.182024 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" event={"ID":"a45e7480-74e1-407a-85ac-03d53fb98a59","Type":"ContainerStarted","Data":"f88f954ba8a4dee2eb0545c306d720f797c22e5f314c152bfc5682ef717c8938"} Apr 17 09:23:46.231386 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:46.231273 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.236208612 podStartE2EDuration="7.231252239s" podCreationTimestamp="2026-04-17 09:23:39 +0000 UTC" firstStartedPulling="2026-04-17 09:23:40.996953237 +0000 UTC m=+153.837990426" lastFinishedPulling="2026-04-17 09:23:45.991996846 +0000 UTC m=+158.833034053" observedRunningTime="2026-04-17 09:23:46.209324894 +0000 UTC m=+159.050362105" watchObservedRunningTime="2026-04-17 09:23:46.231252239 +0000 UTC m=+159.072289452" Apr 17 09:23:46.232299 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:46.232248 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" podStartSLOduration=1.4492495 podStartE2EDuration="3.232237412s" podCreationTimestamp="2026-04-17 09:23:43 +0000 UTC" firstStartedPulling="2026-04-17 09:23:44.206238227 +0000 UTC m=+157.047275416" lastFinishedPulling="2026-04-17 09:23:45.989226135 +0000 UTC m=+158.830263328" observedRunningTime="2026-04-17 09:23:46.230689921 +0000 UTC m=+159.071727133" watchObservedRunningTime="2026-04-17 09:23:46.232237412 +0000 UTC m=+159.073274625" Apr 17 09:23:47.099476 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:47.099447 2569 patch_prober.go:28] interesting pod/image-registry-fcffd4c79-f7kd9 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 17 09:23:47.099627 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:47.099495 2569 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" podUID="5ab3de37-fccc-4ce9-9c52-a0d690b6311a" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 17 09:23:47.187317 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:47.187279 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" event={"ID":"52442855-76a8-4ca3-a454-53fe40b0e7d0","Type":"ContainerStarted","Data":"916b5aa4ab417bb63e16948c735f59923cabe6deac1507beefc8596ff2f9caec"} Apr 17 09:23:47.211624 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:47.211581 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" podStartSLOduration=2.523362408 podStartE2EDuration="6.211566815s" podCreationTimestamp="2026-04-17 09:23:41 +0000 UTC" firstStartedPulling="2026-04-17 09:23:42.300969517 +0000 UTC m=+155.142006706" lastFinishedPulling="2026-04-17 09:23:45.989173923 +0000 UTC m=+158.830211113" observedRunningTime="2026-04-17 09:23:47.208958739 +0000 UTC m=+160.049995963" watchObservedRunningTime="2026-04-17 09:23:47.211566815 +0000 UTC m=+160.052604100" Apr 17 09:23:48.190239 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:48.190213 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:48.949035 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:48.948996 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:23:48.949266 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:48.949066 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:23:48.951313 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:48.951289 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e5c6f5f4-e252-4726-9992-25c01cae77a3-cert\") pod \"ingress-canary-pf795\" (UID: \"e5c6f5f4-e252-4726-9992-25c01cae77a3\") " pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:23:48.951415 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:48.951337 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b171f472-b404-4138-bf91-727a857e3224-metrics-tls\") pod \"dns-default-27fm4\" (UID: \"b171f472-b404-4138-bf91-727a857e3224\") " pod="openshift-dns/dns-default-27fm4" Apr 17 09:23:48.967268 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:48.967246 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-nblxl\"" Apr 17 09:23:48.974983 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:48.974967 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-27fm4" Apr 17 09:23:49.086639 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:49.086619 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-27fm4"] Apr 17 09:23:49.088782 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:49.088748 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb171f472_b404_4138_bf91_727a857e3224.slice/crio-44519df7126c6e6e6b775a9d15be36ce9e764d43c229b7cfc3fd73bbe6205613 WatchSource:0}: Error finding container 44519df7126c6e6e6b775a9d15be36ce9e764d43c229b7cfc3fd73bbe6205613: Status 404 returned error can't find the container with id 44519df7126c6e6e6b775a9d15be36ce9e764d43c229b7cfc3fd73bbe6205613 Apr 17 09:23:49.194052 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:49.194014 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-27fm4" event={"ID":"b171f472-b404-4138-bf91-727a857e3224","Type":"ContainerStarted","Data":"44519df7126c6e6e6b775a9d15be36ce9e764d43c229b7cfc3fd73bbe6205613"} Apr 17 09:23:49.201113 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:49.201056 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-577f9447f9-8p9fd" Apr 17 09:23:50.084341 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.084274 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" podUID="414546fe-e0a7-4aad-bf7e-fa775be0036d" containerName="registry" containerID="cri-o://dbe56c5127783c76b76a8a2c1ffb0ee70ae10f29a71eb4b0ec02a3aff4bb06bc" gracePeriod=30 Apr 17 09:23:50.199525 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.199496 2569 generic.go:358] "Generic (PLEG): container finished" podID="414546fe-e0a7-4aad-bf7e-fa775be0036d" containerID="dbe56c5127783c76b76a8a2c1ffb0ee70ae10f29a71eb4b0ec02a3aff4bb06bc" exitCode=0 Apr 17 09:23:50.199961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.199572 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" event={"ID":"414546fe-e0a7-4aad-bf7e-fa775be0036d","Type":"ContainerDied","Data":"dbe56c5127783c76b76a8a2c1ffb0ee70ae10f29a71eb4b0ec02a3aff4bb06bc"} Apr 17 09:23:50.632476 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.632454 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:23:50.768674 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768649 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/414546fe-e0a7-4aad-bf7e-fa775be0036d-ca-trust-extracted\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.768781 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768689 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-trusted-ca\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.768781 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768721 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-installation-pull-secrets\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.768781 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768751 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-image-registry-private-configuration\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.768998 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768783 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-bound-sa-token\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.768998 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768853 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-certificates\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.768998 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768900 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.768998 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.768959 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vt52\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-kube-api-access-7vt52\") pod \"414546fe-e0a7-4aad-bf7e-fa775be0036d\" (UID: \"414546fe-e0a7-4aad-bf7e-fa775be0036d\") " Apr 17 09:23:50.769396 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.769333 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:23:50.769503 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.769476 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:23:50.771676 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.771628 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:23:50.771676 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.771664 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-kube-api-access-7vt52" (OuterVolumeSpecName: "kube-api-access-7vt52") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "kube-api-access-7vt52". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:23:50.771805 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.771751 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:23:50.771805 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.771775 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:23:50.771949 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.771929 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:23:50.779914 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.779887 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/414546fe-e0a7-4aad-bf7e-fa775be0036d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "414546fe-e0a7-4aad-bf7e-fa775be0036d" (UID: "414546fe-e0a7-4aad-bf7e-fa775be0036d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:23:50.870071 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870049 2569 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-certificates\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:50.870071 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870072 2569 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-registry-tls\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:50.870198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870085 2569 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7vt52\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-kube-api-access-7vt52\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:50.870198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870094 2569 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/414546fe-e0a7-4aad-bf7e-fa775be0036d-ca-trust-extracted\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:50.870198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870104 2569 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/414546fe-e0a7-4aad-bf7e-fa775be0036d-trusted-ca\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:50.870198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870113 2569 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-installation-pull-secrets\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:50.870198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870122 2569 reconciler_common.go:299] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/414546fe-e0a7-4aad-bf7e-fa775be0036d-image-registry-private-configuration\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:50.870198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:50.870130 2569 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/414546fe-e0a7-4aad-bf7e-fa775be0036d-bound-sa-token\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:23:51.204612 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.204513 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-27fm4" event={"ID":"b171f472-b404-4138-bf91-727a857e3224","Type":"ContainerStarted","Data":"90d132532db7495750fcc3c921b6f1089e4082a4f635ec535e4e7d019a664ef0"} Apr 17 09:23:51.204612 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.204552 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-27fm4" event={"ID":"b171f472-b404-4138-bf91-727a857e3224","Type":"ContainerStarted","Data":"b870c09f7ab975dabd249ae65d6ecafa83b372c9b055607b1f78733c684fdffd"} Apr 17 09:23:51.205108 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.204617 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-27fm4" Apr 17 09:23:51.205692 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.205674 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" Apr 17 09:23:51.205692 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.205683 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-75d68f6767-9txgs" event={"ID":"414546fe-e0a7-4aad-bf7e-fa775be0036d","Type":"ContainerDied","Data":"f698f601d7d4f3813728e0633ef14583d5358fb8fe657b9ed916b6f2e79d82fd"} Apr 17 09:23:51.205856 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.205718 2569 scope.go:117] "RemoveContainer" containerID="dbe56c5127783c76b76a8a2c1ffb0ee70ae10f29a71eb4b0ec02a3aff4bb06bc" Apr 17 09:23:51.220181 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.220134 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-27fm4" podStartSLOduration=129.644809363 podStartE2EDuration="2m11.220119225s" podCreationTimestamp="2026-04-17 09:21:40 +0000 UTC" firstStartedPulling="2026-04-17 09:23:49.091086839 +0000 UTC m=+161.932124029" lastFinishedPulling="2026-04-17 09:23:50.66639669 +0000 UTC m=+163.507433891" observedRunningTime="2026-04-17 09:23:51.219561366 +0000 UTC m=+164.060598578" watchObservedRunningTime="2026-04-17 09:23:51.220119225 +0000 UTC m=+164.061156438" Apr 17 09:23:51.231309 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.231285 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-75d68f6767-9txgs"] Apr 17 09:23:51.237159 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.237139 2569 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-75d68f6767-9txgs"] Apr 17 09:23:51.705969 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:51.705938 2569 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="414546fe-e0a7-4aad-bf7e-fa775be0036d" path="/var/lib/kubelet/pods/414546fe-e0a7-4aad-bf7e-fa775be0036d/volumes" Apr 17 09:23:53.702719 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:53.702689 2569 scope.go:117] "RemoveContainer" containerID="eba2cd46dafcad700c8f2567c593d0b9d672409d8151ffd24bc9bf55690d36cc" Apr 17 09:23:54.216175 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.216149 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/2.log" Apr 17 09:23:54.216307 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.216223 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" event={"ID":"ef4e2c76-47f7-42fb-a0e4-d95b609d0101","Type":"ContainerStarted","Data":"439f2123562684ea4547a4d0495fcd304526c80e5d4576f0471f966a48841894"} Apr 17 09:23:54.216483 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.216461 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:23:54.364650 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.364622 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" Apr 17 09:23:54.383668 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.383619 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-9d4b6777b-z9nkj" podStartSLOduration=57.579418054 podStartE2EDuration="59.3836027s" podCreationTimestamp="2026-04-17 09:22:55 +0000 UTC" firstStartedPulling="2026-04-17 09:22:56.418313032 +0000 UTC m=+109.259350221" lastFinishedPulling="2026-04-17 09:22:58.222497678 +0000 UTC m=+111.063534867" observedRunningTime="2026-04-17 09:23:54.243936584 +0000 UTC m=+167.084973798" watchObservedRunningTime="2026-04-17 09:23:54.3836027 +0000 UTC m=+167.224639913" Apr 17 09:23:54.545830 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.545776 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-6bcc868b7-svf8b"] Apr 17 09:23:54.546190 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.546172 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="414546fe-e0a7-4aad-bf7e-fa775be0036d" containerName="registry" Apr 17 09:23:54.546272 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.546193 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="414546fe-e0a7-4aad-bf7e-fa775be0036d" containerName="registry" Apr 17 09:23:54.546272 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.546266 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="414546fe-e0a7-4aad-bf7e-fa775be0036d" containerName="registry" Apr 17 09:23:54.549330 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.549309 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-6bcc868b7-svf8b" Apr 17 09:23:54.551616 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.551598 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 17 09:23:54.551711 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.551634 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 17 09:23:54.551711 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.551643 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-llmwm\"" Apr 17 09:23:54.555767 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.555747 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-6bcc868b7-svf8b"] Apr 17 09:23:54.602874 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.602850 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj46c\" (UniqueName: \"kubernetes.io/projected/a9543ac2-585d-407a-92ca-f7b7437d36f1-kube-api-access-jj46c\") pod \"downloads-6bcc868b7-svf8b\" (UID: \"a9543ac2-585d-407a-92ca-f7b7437d36f1\") " pod="openshift-console/downloads-6bcc868b7-svf8b" Apr 17 09:23:54.704028 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.703999 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jj46c\" (UniqueName: \"kubernetes.io/projected/a9543ac2-585d-407a-92ca-f7b7437d36f1-kube-api-access-jj46c\") pod \"downloads-6bcc868b7-svf8b\" (UID: \"a9543ac2-585d-407a-92ca-f7b7437d36f1\") " pod="openshift-console/downloads-6bcc868b7-svf8b" Apr 17 09:23:54.711299 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.711280 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj46c\" (UniqueName: \"kubernetes.io/projected/a9543ac2-585d-407a-92ca-f7b7437d36f1-kube-api-access-jj46c\") pod \"downloads-6bcc868b7-svf8b\" (UID: \"a9543ac2-585d-407a-92ca-f7b7437d36f1\") " pod="openshift-console/downloads-6bcc868b7-svf8b" Apr 17 09:23:54.858910 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.858854 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-6bcc868b7-svf8b" Apr 17 09:23:54.990007 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:54.989915 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-6bcc868b7-svf8b"] Apr 17 09:23:54.992467 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:54.992442 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9543ac2_585d_407a_92ca_f7b7437d36f1.slice/crio-9b7682d9cab4d136d56571b18ac808216d5d6e63929a2ca4e107689aea27e97b WatchSource:0}: Error finding container 9b7682d9cab4d136d56571b18ac808216d5d6e63929a2ca4e107689aea27e97b: Status 404 returned error can't find the container with id 9b7682d9cab4d136d56571b18ac808216d5d6e63929a2ca4e107689aea27e97b Apr 17 09:23:55.220182 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:55.220110 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-6bcc868b7-svf8b" event={"ID":"a9543ac2-585d-407a-92ca-f7b7437d36f1","Type":"ContainerStarted","Data":"9b7682d9cab4d136d56571b18ac808216d5d6e63929a2ca4e107689aea27e97b"} Apr 17 09:23:55.427681 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:55.427651 2569 patch_prober.go:28] interesting pod/image-registry-fcffd4c79-f7kd9 container/registry namespace/openshift-image-registry: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 17 09:23:55.427843 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:55.427695 2569 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" podUID="5ab3de37-fccc-4ce9-9c52-a0d690b6311a" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 17 09:23:57.099457 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:57.099428 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-fcffd4c79-f7kd9" Apr 17 09:23:57.706228 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:57.706196 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:23:57.709799 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:57.709774 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9szrj\"" Apr 17 09:23:57.716764 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:57.716743 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pf795" Apr 17 09:23:57.852252 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:57.852218 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pf795"] Apr 17 09:23:57.854072 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:23:57.854042 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5c6f5f4_e252_4726_9992_25c01cae77a3.slice/crio-c98d8403ab0c4205c6db0332fb5cc55fe2062c9ea8460ce4415dd648d588bd99 WatchSource:0}: Error finding container c98d8403ab0c4205c6db0332fb5cc55fe2062c9ea8460ce4415dd648d588bd99: Status 404 returned error can't find the container with id c98d8403ab0c4205c6db0332fb5cc55fe2062c9ea8460ce4415dd648d588bd99 Apr 17 09:23:58.230664 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:23:58.230631 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pf795" event={"ID":"e5c6f5f4-e252-4726-9992-25c01cae77a3","Type":"ContainerStarted","Data":"c98d8403ab0c4205c6db0332fb5cc55fe2062c9ea8460ce4415dd648d588bd99"} Apr 17 09:24:00.239086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:00.239044 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pf795" event={"ID":"e5c6f5f4-e252-4726-9992-25c01cae77a3","Type":"ContainerStarted","Data":"b2e2442aa09334cab5c90e3dd1ca968a9f20c91bbbc6187e7f3181f313766421"} Apr 17 09:24:00.255254 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:00.255201 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pf795" podStartSLOduration=138.516037043 podStartE2EDuration="2m20.255183217s" podCreationTimestamp="2026-04-17 09:21:40 +0000 UTC" firstStartedPulling="2026-04-17 09:23:57.856388972 +0000 UTC m=+170.697426162" lastFinishedPulling="2026-04-17 09:23:59.595535145 +0000 UTC m=+172.436572336" observedRunningTime="2026-04-17 09:24:00.253454147 +0000 UTC m=+173.094491362" watchObservedRunningTime="2026-04-17 09:24:00.255183217 +0000 UTC m=+173.096220432" Apr 17 09:24:01.211030 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:01.210998 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-27fm4" Apr 17 09:24:03.498400 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:03.498363 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:24:03.498783 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:03.498414 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:24:04.854876 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.854843 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-5c445f768d-smsjr"] Apr 17 09:24:04.859881 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.859858 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:04.862167 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.862132 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 17 09:24:04.862997 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.862974 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 17 09:24:04.863103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.863024 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-5tsmx\"" Apr 17 09:24:04.863103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.863037 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 17 09:24:04.863103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.863054 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 17 09:24:04.863103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.863066 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 17 09:24:04.875678 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:04.875631 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c445f768d-smsjr"] Apr 17 09:24:05.002023 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.001987 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-oauth-config\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.002189 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.002089 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-service-ca\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.002189 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.002152 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-oauth-serving-cert\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.002189 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.002186 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-console-config\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.002313 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.002217 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg6n5\" (UniqueName: \"kubernetes.io/projected/01aec620-8a18-45b3-8e90-2fa9e535583a-kube-api-access-bg6n5\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.002313 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.002258 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-serving-cert\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.103069 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103031 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-oauth-config\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.103236 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103206 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-service-ca\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.103302 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103251 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-oauth-serving-cert\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.103302 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103275 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-console-config\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.103453 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103427 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bg6n5\" (UniqueName: \"kubernetes.io/projected/01aec620-8a18-45b3-8e90-2fa9e535583a-kube-api-access-bg6n5\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.103515 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103488 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-serving-cert\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.104105 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103991 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-console-config\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.104105 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.103989 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-service-ca\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.104302 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.104133 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-oauth-serving-cert\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.105774 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.105724 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-oauth-config\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.105901 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.105888 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-serving-cert\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.111231 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.111212 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg6n5\" (UniqueName: \"kubernetes.io/projected/01aec620-8a18-45b3-8e90-2fa9e535583a-kube-api-access-bg6n5\") pod \"console-5c445f768d-smsjr\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:05.171218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:05.171162 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:10.546071 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:10.546048 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c445f768d-smsjr"] Apr 17 09:24:10.548418 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:24:10.548393 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01aec620_8a18_45b3_8e90_2fa9e535583a.slice/crio-2b1bad8b63a3ba152b528cf94621c42769c555b77f3ec0ba9becd1e8683a2103 WatchSource:0}: Error finding container 2b1bad8b63a3ba152b528cf94621c42769c555b77f3ec0ba9becd1e8683a2103: Status 404 returned error can't find the container with id 2b1bad8b63a3ba152b528cf94621c42769c555b77f3ec0ba9becd1e8683a2103 Apr 17 09:24:11.278433 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:11.278373 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-6bcc868b7-svf8b" event={"ID":"a9543ac2-585d-407a-92ca-f7b7437d36f1","Type":"ContainerStarted","Data":"65729684db19e219331b8507e9a7e8109089fca2918c6ea02a34f8ddd571d4f5"} Apr 17 09:24:11.278904 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:11.278857 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/downloads-6bcc868b7-svf8b" Apr 17 09:24:11.280757 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:11.280494 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c445f768d-smsjr" event={"ID":"01aec620-8a18-45b3-8e90-2fa9e535583a","Type":"ContainerStarted","Data":"2b1bad8b63a3ba152b528cf94621c42769c555b77f3ec0ba9becd1e8683a2103"} Apr 17 09:24:11.292544 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:11.292508 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-6bcc868b7-svf8b" Apr 17 09:24:11.296455 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:11.296001 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-6bcc868b7-svf8b" podStartSLOduration=1.796312807 podStartE2EDuration="17.295984477s" podCreationTimestamp="2026-04-17 09:23:54 +0000 UTC" firstStartedPulling="2026-04-17 09:23:54.994381562 +0000 UTC m=+167.835418756" lastFinishedPulling="2026-04-17 09:24:10.494053228 +0000 UTC m=+183.335090426" observedRunningTime="2026-04-17 09:24:11.293911304 +0000 UTC m=+184.134948515" watchObservedRunningTime="2026-04-17 09:24:11.295984477 +0000 UTC m=+184.137021690" Apr 17 09:24:15.245268 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.245230 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-7f67678f9c-q8ssl"] Apr 17 09:24:15.259530 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.259501 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7f67678f9c-q8ssl"] Apr 17 09:24:15.259674 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.259618 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.267605 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.267577 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 17 09:24:15.298393 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.298362 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-oauth-serving-cert\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.298523 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.298432 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-service-ca\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.298523 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.298459 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9stf\" (UniqueName: \"kubernetes.io/projected/fe2034cd-2626-49a5-a7ef-25f0d96be762-kube-api-access-t9stf\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.298523 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.298502 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-serving-cert\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.298675 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.298532 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-trusted-ca-bundle\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.298675 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.298570 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-oauth-config\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.298675 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.298593 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-config\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.306410 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.306387 2569 generic.go:358] "Generic (PLEG): container finished" podID="89378780-0d1b-40cd-930c-da0d27751523" containerID="8f18929d00b67b96c922cd27356f0c5055a0b46b88bd16adb289c9bf7a50816e" exitCode=0 Apr 17 09:24:15.306500 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.306471 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" event={"ID":"89378780-0d1b-40cd-930c-da0d27751523","Type":"ContainerDied","Data":"8f18929d00b67b96c922cd27356f0c5055a0b46b88bd16adb289c9bf7a50816e"} Apr 17 09:24:15.306897 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.306874 2569 scope.go:117] "RemoveContainer" containerID="8f18929d00b67b96c922cd27356f0c5055a0b46b88bd16adb289c9bf7a50816e" Apr 17 09:24:15.308200 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.308173 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c445f768d-smsjr" event={"ID":"01aec620-8a18-45b3-8e90-2fa9e535583a","Type":"ContainerStarted","Data":"35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af"} Apr 17 09:24:15.337504 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.337386 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5c445f768d-smsjr" podStartSLOduration=7.673761071 podStartE2EDuration="11.337371453s" podCreationTimestamp="2026-04-17 09:24:04 +0000 UTC" firstStartedPulling="2026-04-17 09:24:10.552888034 +0000 UTC m=+183.393925228" lastFinishedPulling="2026-04-17 09:24:14.216498416 +0000 UTC m=+187.057535610" observedRunningTime="2026-04-17 09:24:15.335981337 +0000 UTC m=+188.177018550" watchObservedRunningTime="2026-04-17 09:24:15.337371453 +0000 UTC m=+188.178408668" Apr 17 09:24:15.400124 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400094 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-oauth-config\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.400235 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400149 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-config\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.400298 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400239 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-oauth-serving-cert\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.400354 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400344 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-service-ca\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.400413 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400366 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-t9stf\" (UniqueName: \"kubernetes.io/projected/fe2034cd-2626-49a5-a7ef-25f0d96be762-kube-api-access-t9stf\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.400469 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400415 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-serving-cert\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.400469 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400436 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-trusted-ca-bundle\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.400883 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.400859 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-config\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.401615 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.401576 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-oauth-serving-cert\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.402035 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.402010 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-trusted-ca-bundle\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.402250 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.402227 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-service-ca\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.403041 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.403022 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-oauth-config\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.403642 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.403621 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-serving-cert\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.408146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.408127 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9stf\" (UniqueName: \"kubernetes.io/projected/fe2034cd-2626-49a5-a7ef-25f0d96be762-kube-api-access-t9stf\") pod \"console-7f67678f9c-q8ssl\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.571923 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.571885 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:15.710330 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:15.710294 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7f67678f9c-q8ssl"] Apr 17 09:24:15.714625 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:24:15.714592 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe2034cd_2626_49a5_a7ef_25f0d96be762.slice/crio-9880de5b36768a36b934d1dbe0aeafaaa2d1ff163fe06d09745b79b7f4739741 WatchSource:0}: Error finding container 9880de5b36768a36b934d1dbe0aeafaaa2d1ff163fe06d09745b79b7f4739741: Status 404 returned error can't find the container with id 9880de5b36768a36b934d1dbe0aeafaaa2d1ff163fe06d09745b79b7f4739741 Apr 17 09:24:16.313380 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:16.313335 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7f67678f9c-q8ssl" event={"ID":"fe2034cd-2626-49a5-a7ef-25f0d96be762","Type":"ContainerStarted","Data":"cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547"} Apr 17 09:24:16.313380 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:16.313377 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7f67678f9c-q8ssl" event={"ID":"fe2034cd-2626-49a5-a7ef-25f0d96be762","Type":"ContainerStarted","Data":"9880de5b36768a36b934d1dbe0aeafaaa2d1ff163fe06d09745b79b7f4739741"} Apr 17 09:24:16.315357 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:16.315328 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-zj5rl" event={"ID":"89378780-0d1b-40cd-930c-da0d27751523","Type":"ContainerStarted","Data":"79307947b308aa74fdc870f4ef04394a073f36f70cd25b1c3677d7f46d8e3172"} Apr 17 09:24:16.330802 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:16.330754 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7f67678f9c-q8ssl" podStartSLOduration=1.330740201 podStartE2EDuration="1.330740201s" podCreationTimestamp="2026-04-17 09:24:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:24:16.328163094 +0000 UTC m=+189.169200309" watchObservedRunningTime="2026-04-17 09:24:16.330740201 +0000 UTC m=+189.171777413" Apr 17 09:24:23.504260 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:23.504223 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:24:23.508170 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:23.508145 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-544766d54-ktmxf" Apr 17 09:24:25.172363 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:25.172327 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:25.172802 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:25.172374 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:25.176948 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:25.176924 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:25.345277 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:25.345255 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:25.572542 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:25.572516 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:25.572542 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:25.572548 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:25.577414 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:25.577392 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:26.146993 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.146966 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_7de2893b-b16a-4458-b369-a1832bca1744/init-config-reloader/0.log" Apr 17 09:24:26.152040 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.152007 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_7de2893b-b16a-4458-b369-a1832bca1744/alertmanager/0.log" Apr 17 09:24:26.311738 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.311713 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_7de2893b-b16a-4458-b369-a1832bca1744/config-reloader/0.log" Apr 17 09:24:26.348301 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.348271 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:24:26.395580 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.395554 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5c445f768d-smsjr"] Apr 17 09:24:26.510405 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.510378 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_7de2893b-b16a-4458-b369-a1832bca1744/kube-rbac-proxy-web/0.log" Apr 17 09:24:26.710704 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.710678 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_7de2893b-b16a-4458-b369-a1832bca1744/kube-rbac-proxy/0.log" Apr 17 09:24:26.911311 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:26.911222 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_7de2893b-b16a-4458-b369-a1832bca1744/kube-rbac-proxy-metric/0.log" Apr 17 09:24:27.110937 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:27.110911 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_7de2893b-b16a-4458-b369-a1832bca1744/prom-label-proxy/0.log" Apr 17 09:24:27.312595 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:27.312572 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-9rpfg_22a71e37-6733-4843-a7c4-2027bbd017e5/cluster-monitoring-operator/0.log" Apr 17 09:24:28.111318 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:28.111297 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_metrics-server-544766d54-ktmxf_a45e7480-74e1-407a-85ac-03d53fb98a59/metrics-server/0.log" Apr 17 09:24:28.512030 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:28.512000 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-fm9gq_4222015d-39a6-4f11-b504-7a3dfc2576cf/init-textfile/0.log" Apr 17 09:24:28.712028 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:28.711986 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-fm9gq_4222015d-39a6-4f11-b504-7a3dfc2576cf/node-exporter/0.log" Apr 17 09:24:28.911599 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:28.911525 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-fm9gq_4222015d-39a6-4f11-b504-7a3dfc2576cf/kube-rbac-proxy/0.log" Apr 17 09:24:30.310793 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:30.310756 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-bttbq_54961548-4e6d-49bc-8671-8265a524aba5/kube-rbac-proxy-main/0.log" Apr 17 09:24:30.511007 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:30.510976 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-bttbq_54961548-4e6d-49bc-8671-8265a524aba5/kube-rbac-proxy-self/0.log" Apr 17 09:24:30.710668 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:30.710580 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-bttbq_54961548-4e6d-49bc-8671-8265a524aba5/openshift-state-metrics/0.log" Apr 17 09:24:32.710448 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:32.710407 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-admission-webhook-57cf98b594-7gvjw_0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d/prometheus-operator-admission-webhook/0.log" Apr 17 09:24:32.910965 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:32.910934 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/thanos-query/0.log" Apr 17 09:24:33.110627 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:33.110593 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy-web/0.log" Apr 17 09:24:33.311287 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:33.311256 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy/0.log" Apr 17 09:24:33.511169 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:33.511146 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/prom-label-proxy/0.log" Apr 17 09:24:33.711124 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:33.711102 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy-rules/0.log" Apr 17 09:24:33.911270 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:33.911200 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy-metrics/0.log" Apr 17 09:24:34.310775 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:34.310748 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/2.log" Apr 17 09:24:34.513066 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:34.513040 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/3.log" Apr 17 09:24:34.711398 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:34.711333 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5c445f768d-smsjr_01aec620-8a18-45b3-8e90-2fa9e535583a/console/0.log" Apr 17 09:24:34.911374 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:34.911348 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7f67678f9c-q8ssl_fe2034cd-2626-49a5-a7ef-25f0d96be762/console/0.log" Apr 17 09:24:35.113242 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:35.113208 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-6bcc868b7-svf8b_a9543ac2-585d-407a-92ca-f7b7437d36f1/download-server/0.log" Apr 17 09:24:52.365711 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.365648 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-5c445f768d-smsjr" podUID="01aec620-8a18-45b3-8e90-2fa9e535583a" containerName="console" containerID="cri-o://35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af" gracePeriod=15 Apr 17 09:24:52.616352 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.616303 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5c445f768d-smsjr_01aec620-8a18-45b3-8e90-2fa9e535583a/console/0.log" Apr 17 09:24:52.616433 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.616359 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:52.713678 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.713651 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-serving-cert\") pod \"01aec620-8a18-45b3-8e90-2fa9e535583a\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " Apr 17 09:24:52.713797 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.713694 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-oauth-serving-cert\") pod \"01aec620-8a18-45b3-8e90-2fa9e535583a\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " Apr 17 09:24:52.713797 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.713727 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-console-config\") pod \"01aec620-8a18-45b3-8e90-2fa9e535583a\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " Apr 17 09:24:52.713797 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.713749 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-service-ca\") pod \"01aec620-8a18-45b3-8e90-2fa9e535583a\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " Apr 17 09:24:52.713937 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.713805 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-oauth-config\") pod \"01aec620-8a18-45b3-8e90-2fa9e535583a\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " Apr 17 09:24:52.713937 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.713866 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg6n5\" (UniqueName: \"kubernetes.io/projected/01aec620-8a18-45b3-8e90-2fa9e535583a-kube-api-access-bg6n5\") pod \"01aec620-8a18-45b3-8e90-2fa9e535583a\" (UID: \"01aec620-8a18-45b3-8e90-2fa9e535583a\") " Apr 17 09:24:52.714169 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.714142 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "01aec620-8a18-45b3-8e90-2fa9e535583a" (UID: "01aec620-8a18-45b3-8e90-2fa9e535583a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:24:52.714270 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.714165 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-console-config" (OuterVolumeSpecName: "console-config") pod "01aec620-8a18-45b3-8e90-2fa9e535583a" (UID: "01aec620-8a18-45b3-8e90-2fa9e535583a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:24:52.714270 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.714175 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-service-ca" (OuterVolumeSpecName: "service-ca") pod "01aec620-8a18-45b3-8e90-2fa9e535583a" (UID: "01aec620-8a18-45b3-8e90-2fa9e535583a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:24:52.715869 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.715812 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "01aec620-8a18-45b3-8e90-2fa9e535583a" (UID: "01aec620-8a18-45b3-8e90-2fa9e535583a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:24:52.715958 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.715897 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "01aec620-8a18-45b3-8e90-2fa9e535583a" (UID: "01aec620-8a18-45b3-8e90-2fa9e535583a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:24:52.715996 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.715977 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01aec620-8a18-45b3-8e90-2fa9e535583a-kube-api-access-bg6n5" (OuterVolumeSpecName: "kube-api-access-bg6n5") pod "01aec620-8a18-45b3-8e90-2fa9e535583a" (UID: "01aec620-8a18-45b3-8e90-2fa9e535583a"). InnerVolumeSpecName "kube-api-access-bg6n5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:24:52.815181 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.815157 2569 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-oauth-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:24:52.815181 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.815179 2569 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bg6n5\" (UniqueName: \"kubernetes.io/projected/01aec620-8a18-45b3-8e90-2fa9e535583a-kube-api-access-bg6n5\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:24:52.815308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.815188 2569 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/01aec620-8a18-45b3-8e90-2fa9e535583a-console-serving-cert\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:24:52.815308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.815197 2569 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-oauth-serving-cert\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:24:52.815308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.815206 2569 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-console-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:24:52.815308 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:52.815214 2569 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01aec620-8a18-45b3-8e90-2fa9e535583a-service-ca\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:24:53.430266 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.430242 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5c445f768d-smsjr_01aec620-8a18-45b3-8e90-2fa9e535583a/console/0.log" Apr 17 09:24:53.430635 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.430276 2569 generic.go:358] "Generic (PLEG): container finished" podID="01aec620-8a18-45b3-8e90-2fa9e535583a" containerID="35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af" exitCode=2 Apr 17 09:24:53.430635 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.430300 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c445f768d-smsjr" event={"ID":"01aec620-8a18-45b3-8e90-2fa9e535583a","Type":"ContainerDied","Data":"35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af"} Apr 17 09:24:53.430635 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.430320 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c445f768d-smsjr" event={"ID":"01aec620-8a18-45b3-8e90-2fa9e535583a","Type":"ContainerDied","Data":"2b1bad8b63a3ba152b528cf94621c42769c555b77f3ec0ba9becd1e8683a2103"} Apr 17 09:24:53.430635 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.430334 2569 scope.go:117] "RemoveContainer" containerID="35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af" Apr 17 09:24:53.430635 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.430337 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c445f768d-smsjr" Apr 17 09:24:53.437977 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.437960 2569 scope.go:117] "RemoveContainer" containerID="35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af" Apr 17 09:24:53.438223 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:24:53.438204 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af\": container with ID starting with 35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af not found: ID does not exist" containerID="35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af" Apr 17 09:24:53.438276 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.438231 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af"} err="failed to get container status \"35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af\": rpc error: code = NotFound desc = could not find container \"35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af\": container with ID starting with 35d36ea42cd26e77357c72e33ccd9cca11a28c96926164f0f44b2e45d8a800af not found: ID does not exist" Apr 17 09:24:53.449353 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.449333 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5c445f768d-smsjr"] Apr 17 09:24:53.452862 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.452840 2569 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5c445f768d-smsjr"] Apr 17 09:24:53.706791 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:53.706729 2569 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01aec620-8a18-45b3-8e90-2fa9e535583a" path="/var/lib/kubelet/pods/01aec620-8a18-45b3-8e90-2fa9e535583a/volumes" Apr 17 09:24:59.150806 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.150769 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:24:59.151246 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.151203 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="alertmanager" containerID="cri-o://a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9" gracePeriod=120 Apr 17 09:24:59.151304 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.151265 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-metric" containerID="cri-o://01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30" gracePeriod=120 Apr 17 09:24:59.151361 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.151297 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-web" containerID="cri-o://49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7" gracePeriod=120 Apr 17 09:24:59.151361 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.151313 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy" containerID="cri-o://ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5" gracePeriod=120 Apr 17 09:24:59.151456 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.151354 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="prom-label-proxy" containerID="cri-o://83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e" gracePeriod=120 Apr 17 09:24:59.151456 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.151329 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="config-reloader" containerID="cri-o://5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683" gracePeriod=120 Apr 17 09:24:59.450566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450498 2569 generic.go:358] "Generic (PLEG): container finished" podID="7de2893b-b16a-4458-b369-a1832bca1744" containerID="83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e" exitCode=0 Apr 17 09:24:59.450566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450522 2569 generic.go:358] "Generic (PLEG): container finished" podID="7de2893b-b16a-4458-b369-a1832bca1744" containerID="ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5" exitCode=0 Apr 17 09:24:59.450566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450529 2569 generic.go:358] "Generic (PLEG): container finished" podID="7de2893b-b16a-4458-b369-a1832bca1744" containerID="5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683" exitCode=0 Apr 17 09:24:59.450566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450535 2569 generic.go:358] "Generic (PLEG): container finished" podID="7de2893b-b16a-4458-b369-a1832bca1744" containerID="a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9" exitCode=0 Apr 17 09:24:59.450751 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450561 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e"} Apr 17 09:24:59.450751 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450590 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5"} Apr 17 09:24:59.450751 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450601 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683"} Apr 17 09:24:59.450751 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:24:59.450609 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9"} Apr 17 09:25:00.390198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.390178 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.456383 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.456349 2569 generic.go:358] "Generic (PLEG): container finished" podID="7de2893b-b16a-4458-b369-a1832bca1744" containerID="01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30" exitCode=0 Apr 17 09:25:00.456383 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.456382 2569 generic.go:358] "Generic (PLEG): container finished" podID="7de2893b-b16a-4458-b369-a1832bca1744" containerID="49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7" exitCode=0 Apr 17 09:25:00.456557 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.456420 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30"} Apr 17 09:25:00.456557 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.456458 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7"} Apr 17 09:25:00.456557 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.456470 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.456557 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.456481 2569 scope.go:117] "RemoveContainer" containerID="83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e" Apr 17 09:25:00.456557 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.456471 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"7de2893b-b16a-4458-b369-a1832bca1744","Type":"ContainerDied","Data":"98f9ce50fbaa76f81a6f7d78426a3223607f17ce40d8a5edc12f44c4196bf722"} Apr 17 09:25:00.463660 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.463646 2569 scope.go:117] "RemoveContainer" containerID="01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30" Apr 17 09:25:00.470237 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.470223 2569 scope.go:117] "RemoveContainer" containerID="ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5" Apr 17 09:25:00.476263 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.476247 2569 scope.go:117] "RemoveContainer" containerID="49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7" Apr 17 09:25:00.481945 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.481931 2569 scope.go:117] "RemoveContainer" containerID="5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683" Apr 17 09:25:00.487721 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.487706 2569 scope.go:117] "RemoveContainer" containerID="a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9" Apr 17 09:25:00.493643 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.493626 2569 scope.go:117] "RemoveContainer" containerID="8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310" Apr 17 09:25:00.499243 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.499225 2569 scope.go:117] "RemoveContainer" containerID="83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e" Apr 17 09:25:00.499457 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:00.499441 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e\": container with ID starting with 83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e not found: ID does not exist" containerID="83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e" Apr 17 09:25:00.499512 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.499464 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e"} err="failed to get container status \"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e\": rpc error: code = NotFound desc = could not find container \"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e\": container with ID starting with 83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e not found: ID does not exist" Apr 17 09:25:00.499512 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.499477 2569 scope.go:117] "RemoveContainer" containerID="01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30" Apr 17 09:25:00.499656 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:00.499640 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30\": container with ID starting with 01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30 not found: ID does not exist" containerID="01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30" Apr 17 09:25:00.499693 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.499663 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30"} err="failed to get container status \"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30\": rpc error: code = NotFound desc = could not find container \"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30\": container with ID starting with 01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30 not found: ID does not exist" Apr 17 09:25:00.499693 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.499679 2569 scope.go:117] "RemoveContainer" containerID="ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5" Apr 17 09:25:00.499966 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:00.499951 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5\": container with ID starting with ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5 not found: ID does not exist" containerID="ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5" Apr 17 09:25:00.500023 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.499969 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5"} err="failed to get container status \"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5\": rpc error: code = NotFound desc = could not find container \"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5\": container with ID starting with ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5 not found: ID does not exist" Apr 17 09:25:00.500023 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.499981 2569 scope.go:117] "RemoveContainer" containerID="49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7" Apr 17 09:25:00.500185 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:00.500171 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7\": container with ID starting with 49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7 not found: ID does not exist" containerID="49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7" Apr 17 09:25:00.500218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500191 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7"} err="failed to get container status \"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7\": rpc error: code = NotFound desc = could not find container \"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7\": container with ID starting with 49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7 not found: ID does not exist" Apr 17 09:25:00.500218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500210 2569 scope.go:117] "RemoveContainer" containerID="5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683" Apr 17 09:25:00.500415 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:00.500402 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683\": container with ID starting with 5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683 not found: ID does not exist" containerID="5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683" Apr 17 09:25:00.500451 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500417 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683"} err="failed to get container status \"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683\": rpc error: code = NotFound desc = could not find container \"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683\": container with ID starting with 5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683 not found: ID does not exist" Apr 17 09:25:00.500451 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500443 2569 scope.go:117] "RemoveContainer" containerID="a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9" Apr 17 09:25:00.500611 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:00.500597 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9\": container with ID starting with a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9 not found: ID does not exist" containerID="a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9" Apr 17 09:25:00.500653 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500615 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9"} err="failed to get container status \"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9\": rpc error: code = NotFound desc = could not find container \"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9\": container with ID starting with a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9 not found: ID does not exist" Apr 17 09:25:00.500653 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500627 2569 scope.go:117] "RemoveContainer" containerID="8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310" Apr 17 09:25:00.500798 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:00.500785 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310\": container with ID starting with 8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310 not found: ID does not exist" containerID="8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310" Apr 17 09:25:00.500884 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500801 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310"} err="failed to get container status \"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310\": rpc error: code = NotFound desc = could not find container \"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310\": container with ID starting with 8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310 not found: ID does not exist" Apr 17 09:25:00.500884 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.500830 2569 scope.go:117] "RemoveContainer" containerID="83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e" Apr 17 09:25:00.501045 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501030 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e"} err="failed to get container status \"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e\": rpc error: code = NotFound desc = could not find container \"83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e\": container with ID starting with 83a796b688d71d54c66768be69fa8042a1efcfb8c637318db55f61d19bfa473e not found: ID does not exist" Apr 17 09:25:00.501045 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501045 2569 scope.go:117] "RemoveContainer" containerID="01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30" Apr 17 09:25:00.501249 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501229 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30"} err="failed to get container status \"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30\": rpc error: code = NotFound desc = could not find container \"01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30\": container with ID starting with 01015c7797f8144f12e1fb474a54ec8f2ca519897fe2da2d78ac29c981204d30 not found: ID does not exist" Apr 17 09:25:00.501290 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501251 2569 scope.go:117] "RemoveContainer" containerID="ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5" Apr 17 09:25:00.501438 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501422 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5"} err="failed to get container status \"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5\": rpc error: code = NotFound desc = could not find container \"ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5\": container with ID starting with ced6aff5b13e9fb401aa090399fdea3e790a3785fe16c75dffaa832865aa66f5 not found: ID does not exist" Apr 17 09:25:00.501484 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501439 2569 scope.go:117] "RemoveContainer" containerID="49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7" Apr 17 09:25:00.501613 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501599 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7"} err="failed to get container status \"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7\": rpc error: code = NotFound desc = could not find container \"49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7\": container with ID starting with 49b3a3c48bf6601c6985d3692790d090e148b268a5eff8de3576dab90dbaa2e7 not found: ID does not exist" Apr 17 09:25:00.501613 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501613 2569 scope.go:117] "RemoveContainer" containerID="5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683" Apr 17 09:25:00.501777 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501759 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683"} err="failed to get container status \"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683\": rpc error: code = NotFound desc = could not find container \"5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683\": container with ID starting with 5d90157f2fa9e716110350cb4533027e0020356cf4e04968b44f35d151698683 not found: ID does not exist" Apr 17 09:25:00.501849 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501778 2569 scope.go:117] "RemoveContainer" containerID="a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9" Apr 17 09:25:00.501972 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501952 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9"} err="failed to get container status \"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9\": rpc error: code = NotFound desc = could not find container \"a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9\": container with ID starting with a28dfa6eb4d514fc86b90fa64a837c9d6738dee09f3277f5d9251c371618dab9 not found: ID does not exist" Apr 17 09:25:00.502040 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.501974 2569 scope.go:117] "RemoveContainer" containerID="8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310" Apr 17 09:25:00.502175 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.502157 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310"} err="failed to get container status \"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310\": rpc error: code = NotFound desc = could not find container \"8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310\": container with ID starting with 8f28b5ac6712c8bdc7002695697cd3128861685979de1282391e881eecdb7310 not found: ID does not exist" Apr 17 09:25:00.570766 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.570746 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8wbx\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-kube-api-access-n8wbx\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.570859 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.570777 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-metrics-client-ca\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.570859 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.570799 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.570859 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.570836 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.570979 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.570891 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-web\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.570979 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.570949 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571077 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.570979 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-web-config\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571127 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571093 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-tls-assets\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571180 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571136 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-metric\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571180 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571167 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-config-volume\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571277 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571172 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:25:00.571277 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571207 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle" (OuterVolumeSpecName: "alertmanager-trusted-ca-bundle") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "alertmanager-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:25:00.571277 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571211 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-cluster-tls-config\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571427 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571279 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-main-db\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571427 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571305 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-config-out\") pod \"7de2893b-b16a-4458-b369-a1832bca1744\" (UID: \"7de2893b-b16a-4458-b369-a1832bca1744\") " Apr 17 09:25:00.571611 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571590 2569 reconciler_common.go:299] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-metrics-client-ca\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.571668 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571620 2569 reconciler_common.go:299] "Volume detached for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-trusted-ca-bundle\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.571848 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.571799 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-main-db" (OuterVolumeSpecName: "alertmanager-main-db") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "alertmanager-main-db". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:25:00.573831 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.573785 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:00.574191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.574137 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-web" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-web") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-web". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:00.574191 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.574172 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-metric" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-metric") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-metric". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:00.574334 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.574278 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-kube-api-access-n8wbx" (OuterVolumeSpecName: "kube-api-access-n8wbx") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "kube-api-access-n8wbx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:25:00.574572 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.574544 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls" (OuterVolumeSpecName: "secret-alertmanager-main-tls") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "secret-alertmanager-main-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:00.575052 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.575024 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:25:00.575278 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.575258 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-config-out" (OuterVolumeSpecName: "config-out") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:25:00.575526 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.575507 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-config-volume" (OuterVolumeSpecName: "config-volume") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:00.577705 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.577681 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-cluster-tls-config" (OuterVolumeSpecName: "cluster-tls-config") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "cluster-tls-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:00.584106 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.584086 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-web-config" (OuterVolumeSpecName: "web-config") pod "7de2893b-b16a-4458-b369-a1832bca1744" (UID: "7de2893b-b16a-4458-b369-a1832bca1744"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:00.672276 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672256 2569 reconciler_common.go:299] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-tls-assets\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672280 2569 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-metric\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672294 2569 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-config-volume\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672307 2569 reconciler_common.go:299] "Volume detached for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-cluster-tls-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672322 2569 reconciler_common.go:299] "Volume detached for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-alertmanager-main-db\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672335 2569 reconciler_common.go:299] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7de2893b-b16a-4458-b369-a1832bca1744-config-out\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672347 2569 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-n8wbx\" (UniqueName: \"kubernetes.io/projected/7de2893b-b16a-4458-b369-a1832bca1744-kube-api-access-n8wbx\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672360 2569 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672586 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672375 2569 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-kube-rbac-proxy-web\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672586 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672389 2569 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-secret-alertmanager-main-tls\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.672586 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.672403 2569 reconciler_common.go:299] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7de2893b-b16a-4458-b369-a1832bca1744-web-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:00.780479 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.780449 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:25:00.784875 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.784854 2569 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:25:00.808010 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.807986 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:25:00.808286 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808274 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="config-reloader" Apr 17 09:25:00.808327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808288 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="config-reloader" Apr 17 09:25:00.808327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808300 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="init-config-reloader" Apr 17 09:25:00.808327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808309 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="init-config-reloader" Apr 17 09:25:00.808327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808322 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="prom-label-proxy" Apr 17 09:25:00.808327 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808327 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="prom-label-proxy" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808338 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-metric" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808343 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-metric" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808350 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-web" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808355 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-web" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808360 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="01aec620-8a18-45b3-8e90-2fa9e535583a" containerName="console" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808365 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="01aec620-8a18-45b3-8e90-2fa9e535583a" containerName="console" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808371 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808376 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808387 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="alertmanager" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808394 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="alertmanager" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808437 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="prom-label-proxy" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808445 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-web" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808452 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy-metric" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808460 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="kube-rbac-proxy" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808469 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="01aec620-8a18-45b3-8e90-2fa9e535583a" containerName="console" Apr 17 09:25:00.808473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808479 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="alertmanager" Apr 17 09:25:00.808906 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.808485 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="7de2893b-b16a-4458-b369-a1832bca1744" containerName="config-reloader" Apr 17 09:25:00.813565 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.813550 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.815862 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.815843 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Apr 17 09:25:00.816006 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.815953 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Apr 17 09:25:00.816090 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.816066 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Apr 17 09:25:00.816148 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.816100 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Apr 17 09:25:00.816223 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.816145 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Apr 17 09:25:00.816223 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.816210 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Apr 17 09:25:00.816346 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.816324 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-g6hlq\"" Apr 17 09:25:00.816404 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.816370 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Apr 17 09:25:00.816404 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.816370 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Apr 17 09:25:00.821894 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.821877 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Apr 17 09:25:00.822623 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.822605 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:25:00.975938 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.975914 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/ab8a248c-4463-4b8d-9ba6-1b2063937460-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976059 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.975951 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976059 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.975971 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976065 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-config-volume\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976104 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976134 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ab8a248c-4463-4b8d-9ba6-1b2063937460-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976300 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976170 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976300 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976197 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976300 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976221 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab8a248c-4463-4b8d-9ba6-1b2063937460-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976300 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976271 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ab8a248c-4463-4b8d-9ba6-1b2063937460-config-out\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976300 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976294 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x5fk\" (UniqueName: \"kubernetes.io/projected/ab8a248c-4463-4b8d-9ba6-1b2063937460-kube-api-access-5x5fk\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976442 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976316 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ab8a248c-4463-4b8d-9ba6-1b2063937460-tls-assets\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:00.976442 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:00.976334 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-web-config\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077122 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077153 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-config-volume\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077173 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077191 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ab8a248c-4463-4b8d-9ba6-1b2063937460-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077212 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077236 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077262 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab8a248c-4463-4b8d-9ba6-1b2063937460-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077289 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ab8a248c-4463-4b8d-9ba6-1b2063937460-config-out\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077313 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5x5fk\" (UniqueName: \"kubernetes.io/projected/ab8a248c-4463-4b8d-9ba6-1b2063937460-kube-api-access-5x5fk\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077342 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ab8a248c-4463-4b8d-9ba6-1b2063937460-tls-assets\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077367 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077364 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-web-config\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077843 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077392 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/ab8a248c-4463-4b8d-9ba6-1b2063937460-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077843 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077446 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.077982 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.077962 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ab8a248c-4463-4b8d-9ba6-1b2063937460-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.079811 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.079700 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.079940 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.079915 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ab8a248c-4463-4b8d-9ba6-1b2063937460-config-out\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.080113 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.080083 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.080290 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.080232 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.080379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.080330 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/ab8a248c-4463-4b8d-9ba6-1b2063937460-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.080379 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.080341 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-config-volume\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.080621 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.080597 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ab8a248c-4463-4b8d-9ba6-1b2063937460-tls-assets\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.080956 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.080934 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab8a248c-4463-4b8d-9ba6-1b2063937460-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.081036 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.081021 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.081487 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.081467 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.081861 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.081844 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ab8a248c-4463-4b8d-9ba6-1b2063937460-web-config\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.088174 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.088155 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x5fk\" (UniqueName: \"kubernetes.io/projected/ab8a248c-4463-4b8d-9ba6-1b2063937460-kube-api-access-5x5fk\") pod \"alertmanager-main-0\" (UID: \"ab8a248c-4463-4b8d-9ba6-1b2063937460\") " pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.123039 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.123018 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 17 09:25:01.241436 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.241396 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 17 09:25:01.244119 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:25:01.244079 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab8a248c_4463_4b8d_9ba6_1b2063937460.slice/crio-991a30523729c465d4894ff31d35ff65f3da33b20d5913fc6782a31861973c55 WatchSource:0}: Error finding container 991a30523729c465d4894ff31d35ff65f3da33b20d5913fc6782a31861973c55: Status 404 returned error can't find the container with id 991a30523729c465d4894ff31d35ff65f3da33b20d5913fc6782a31861973c55 Apr 17 09:25:01.460922 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.460897 2569 generic.go:358] "Generic (PLEG): container finished" podID="ab8a248c-4463-4b8d-9ba6-1b2063937460" containerID="256c84e82935d4885721f14cf195e3590786f98dabee3eb0eff14dca7b663e43" exitCode=0 Apr 17 09:25:01.461271 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.460988 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerDied","Data":"256c84e82935d4885721f14cf195e3590786f98dabee3eb0eff14dca7b663e43"} Apr 17 09:25:01.461271 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.461022 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerStarted","Data":"991a30523729c465d4894ff31d35ff65f3da33b20d5913fc6782a31861973c55"} Apr 17 09:25:01.707296 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:01.707274 2569 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7de2893b-b16a-4458-b369-a1832bca1744" path="/var/lib/kubelet/pods/7de2893b-b16a-4458-b369-a1832bca1744/volumes" Apr 17 09:25:02.467362 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:02.467330 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerStarted","Data":"1b45bf927e2576b2590b2d8de781b26a6bb665537b2ba5b78f8a7c83c450b2ff"} Apr 17 09:25:02.467362 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:02.467366 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerStarted","Data":"dc1edac505189b078d8a1469340897b4901f2e43b1530080819dad129fe1838d"} Apr 17 09:25:02.467749 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:02.467375 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerStarted","Data":"c068c768e8dec7096411ab72e8db81ae8896d705edf6285bbfd2c13589a0b783"} Apr 17 09:25:02.467749 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:02.467384 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerStarted","Data":"bfc3694b87640558ee9d27818076b162a4e6857f9bd1bfa66ce472e16d663a84"} Apr 17 09:25:02.467749 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:02.467391 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerStarted","Data":"ab64d569fc8d3dcf46d947faad0702e7ca5d13fb3f9ca256cb72104d321f55ae"} Apr 17 09:25:02.467749 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:02.467401 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"ab8a248c-4463-4b8d-9ba6-1b2063937460","Type":"ContainerStarted","Data":"651838295d92d34705837daf577263ad2da47979e9d1f7bca62791f5a2ac6ab3"} Apr 17 09:25:02.493003 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:02.492920 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.492879415 podStartE2EDuration="2.492879415s" podCreationTimestamp="2026-04-17 09:25:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:25:02.491145487 +0000 UTC m=+235.332182699" watchObservedRunningTime="2026-04-17 09:25:02.492879415 +0000 UTC m=+235.333916626" Apr 17 09:25:03.173229 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.173194 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/telemeter-client-9798999bf-dvwvs"] Apr 17 09:25:03.176808 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.176788 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.179319 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.179295 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemeter-client-serving-certs-ca-bundle\"" Apr 17 09:25:03.179443 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.179326 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-tls\"" Apr 17 09:25:03.179443 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.179406 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-dockercfg-2n96d\"" Apr 17 09:25:03.179443 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.179428 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client\"" Apr 17 09:25:03.179575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.179442 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-kube-rbac-proxy-config\"" Apr 17 09:25:03.179575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.179570 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"federate-client-certs\"" Apr 17 09:25:03.183551 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.183524 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemeter-trusted-ca-bundle-8i12ta5c71j38\"" Apr 17 09:25:03.185647 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.185629 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-9798999bf-dvwvs"] Apr 17 09:25:03.294685 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294651 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-telemeter-trusted-ca-bundle\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.294685 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294687 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-serving-certs-ca-bundle\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.294920 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294715 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrbjs\" (UniqueName: \"kubernetes.io/projected/bb3fcb85-ac31-47ee-b6e7-614532698db2-kube-api-access-xrbjs\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.294920 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294762 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-metrics-client-ca\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.294920 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294809 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-federate-client-tls\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.294920 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294911 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-telemeter-client-tls\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.295061 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294948 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.295061 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.294974 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-secret-telemeter-client\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395736 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395707 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-telemeter-trusted-ca-bundle\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395736 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395737 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-serving-certs-ca-bundle\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395757 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xrbjs\" (UniqueName: \"kubernetes.io/projected/bb3fcb85-ac31-47ee-b6e7-614532698db2-kube-api-access-xrbjs\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395782 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-metrics-client-ca\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395803 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-federate-client-tls\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395872 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-telemeter-client-tls\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395917 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.395961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.395951 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-secret-telemeter-client\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.396543 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.396515 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-serving-certs-ca-bundle\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.396635 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.396589 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-metrics-client-ca\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.396692 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.396628 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb3fcb85-ac31-47ee-b6e7-614532698db2-telemeter-trusted-ca-bundle\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.398254 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.398229 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.398429 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.398410 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-secret-telemeter-client\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.398575 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.398556 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-telemeter-client-tls\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.398637 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.398589 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/bb3fcb85-ac31-47ee-b6e7-614532698db2-federate-client-tls\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.402746 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.402730 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrbjs\" (UniqueName: \"kubernetes.io/projected/bb3fcb85-ac31-47ee-b6e7-614532698db2-kube-api-access-xrbjs\") pod \"telemeter-client-9798999bf-dvwvs\" (UID: \"bb3fcb85-ac31-47ee-b6e7-614532698db2\") " pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.488277 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.488203 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" Apr 17 09:25:03.605043 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:03.605019 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-9798999bf-dvwvs"] Apr 17 09:25:03.606472 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:25:03.606446 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb3fcb85_ac31_47ee_b6e7_614532698db2.slice/crio-c02cc0bffe745edfabb74ca630819136ff59d5f7aa494c0a0b1c8b5d0357bb7a WatchSource:0}: Error finding container c02cc0bffe745edfabb74ca630819136ff59d5f7aa494c0a0b1c8b5d0357bb7a: Status 404 returned error can't find the container with id c02cc0bffe745edfabb74ca630819136ff59d5f7aa494c0a0b1c8b5d0357bb7a Apr 17 09:25:04.475100 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:04.475056 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" event={"ID":"bb3fcb85-ac31-47ee-b6e7-614532698db2","Type":"ContainerStarted","Data":"c02cc0bffe745edfabb74ca630819136ff59d5f7aa494c0a0b1c8b5d0357bb7a"} Apr 17 09:25:06.487002 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:06.486966 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" event={"ID":"bb3fcb85-ac31-47ee-b6e7-614532698db2","Type":"ContainerStarted","Data":"df57751ffc7b9743b29365e24e3572f205f0199ba5519d188c36d40d5cc0ba63"} Apr 17 09:25:06.487002 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:06.487003 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" event={"ID":"bb3fcb85-ac31-47ee-b6e7-614532698db2","Type":"ContainerStarted","Data":"e016b34072d0624766a429134be65e527f4a3ba7f3ffbe270073447a6613297e"} Apr 17 09:25:06.487002 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:06.487012 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" event={"ID":"bb3fcb85-ac31-47ee-b6e7-614532698db2","Type":"ContainerStarted","Data":"a6f6540d8b1cfa0adfa043c40a0c05c6eaef71aa91e60baabbc07d133e51da68"} Apr 17 09:25:06.507077 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:06.507027 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/telemeter-client-9798999bf-dvwvs" podStartSLOduration=1.679878075 podStartE2EDuration="3.507012082s" podCreationTimestamp="2026-04-17 09:25:03 +0000 UTC" firstStartedPulling="2026-04-17 09:25:03.608380659 +0000 UTC m=+236.449417848" lastFinishedPulling="2026-04-17 09:25:05.435514662 +0000 UTC m=+238.276551855" observedRunningTime="2026-04-17 09:25:06.506342088 +0000 UTC m=+239.347379297" watchObservedRunningTime="2026-04-17 09:25:06.507012082 +0000 UTC m=+239.348049292" Apr 17 09:25:07.125736 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.125706 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-fbd67bd86-86j2d"] Apr 17 09:25:07.129422 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.129399 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.140654 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.140631 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-fbd67bd86-86j2d"] Apr 17 09:25:07.226111 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.226080 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-serving-cert\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.226111 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.226113 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-config\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.226326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.226135 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99c8c\" (UniqueName: \"kubernetes.io/projected/dacde601-d4d3-4635-bf77-62a3cd0c1499-kube-api-access-99c8c\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.226326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.226170 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-oauth-serving-cert\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.226326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.226215 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-service-ca\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.226326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.226236 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-oauth-config\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.226326 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.226260 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-trusted-ca-bundle\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.326651 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.326619 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-oauth-serving-cert\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.326837 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.326657 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-service-ca\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.326837 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.326710 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-oauth-config\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.326837 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.326748 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-trusted-ca-bundle\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.326837 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.326785 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-serving-cert\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.326837 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.326835 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-config\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.327085 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.326880 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-99c8c\" (UniqueName: \"kubernetes.io/projected/dacde601-d4d3-4635-bf77-62a3cd0c1499-kube-api-access-99c8c\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.327550 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.327521 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-oauth-serving-cert\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.327550 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.327544 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-config\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.327719 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.327614 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-service-ca\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.327719 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.327653 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-trusted-ca-bundle\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.329236 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.329215 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-oauth-config\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.329368 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.329351 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-serving-cert\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.333583 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.333559 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-99c8c\" (UniqueName: \"kubernetes.io/projected/dacde601-d4d3-4635-bf77-62a3cd0c1499-kube-api-access-99c8c\") pod \"console-fbd67bd86-86j2d\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.439959 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.439888 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:07.768809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:07.768787 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-fbd67bd86-86j2d"] Apr 17 09:25:07.770850 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:25:07.770801 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddacde601_d4d3_4635_bf77_62a3cd0c1499.slice/crio-4333d05a468d83d6d393554438c56430b20e85046efcce1ba49e39f5fcebaee4 WatchSource:0}: Error finding container 4333d05a468d83d6d393554438c56430b20e85046efcce1ba49e39f5fcebaee4: Status 404 returned error can't find the container with id 4333d05a468d83d6d393554438c56430b20e85046efcce1ba49e39f5fcebaee4 Apr 17 09:25:08.494618 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:08.494587 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-fbd67bd86-86j2d" event={"ID":"dacde601-d4d3-4635-bf77-62a3cd0c1499","Type":"ContainerStarted","Data":"ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d"} Apr 17 09:25:08.494618 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:08.494621 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-fbd67bd86-86j2d" event={"ID":"dacde601-d4d3-4635-bf77-62a3cd0c1499","Type":"ContainerStarted","Data":"4333d05a468d83d6d393554438c56430b20e85046efcce1ba49e39f5fcebaee4"} Apr 17 09:25:08.515855 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:08.515797 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-fbd67bd86-86j2d" podStartSLOduration=1.51578408 podStartE2EDuration="1.51578408s" podCreationTimestamp="2026-04-17 09:25:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:25:08.513949021 +0000 UTC m=+241.354986232" watchObservedRunningTime="2026-04-17 09:25:08.51578408 +0000 UTC m=+241.356821291" Apr 17 09:25:17.441055 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:17.440965 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:17.441055 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:17.441011 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:17.445416 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:17.445394 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:17.525566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:17.525543 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:25:17.569773 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:17.569739 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7f67678f9c-q8ssl"] Apr 17 09:25:42.590910 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.590840 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-7f67678f9c-q8ssl" podUID="fe2034cd-2626-49a5-a7ef-25f0d96be762" containerName="console" containerID="cri-o://cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547" gracePeriod=15 Apr 17 09:25:42.826057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.826038 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7f67678f9c-q8ssl_fe2034cd-2626-49a5-a7ef-25f0d96be762/console/0.log" Apr 17 09:25:42.826159 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.826095 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:25:42.886789 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.886715 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-config\") pod \"fe2034cd-2626-49a5-a7ef-25f0d96be762\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " Apr 17 09:25:42.886789 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.886759 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9stf\" (UniqueName: \"kubernetes.io/projected/fe2034cd-2626-49a5-a7ef-25f0d96be762-kube-api-access-t9stf\") pod \"fe2034cd-2626-49a5-a7ef-25f0d96be762\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " Apr 17 09:25:42.886789 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.886787 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-service-ca\") pod \"fe2034cd-2626-49a5-a7ef-25f0d96be762\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " Apr 17 09:25:42.887038 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.886807 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-serving-cert\") pod \"fe2034cd-2626-49a5-a7ef-25f0d96be762\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " Apr 17 09:25:42.887038 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.886833 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-trusted-ca-bundle\") pod \"fe2034cd-2626-49a5-a7ef-25f0d96be762\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " Apr 17 09:25:42.887038 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.886909 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-oauth-serving-cert\") pod \"fe2034cd-2626-49a5-a7ef-25f0d96be762\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " Apr 17 09:25:42.887038 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.886972 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-oauth-config\") pod \"fe2034cd-2626-49a5-a7ef-25f0d96be762\" (UID: \"fe2034cd-2626-49a5-a7ef-25f0d96be762\") " Apr 17 09:25:42.887245 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.887169 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-config" (OuterVolumeSpecName: "console-config") pod "fe2034cd-2626-49a5-a7ef-25f0d96be762" (UID: "fe2034cd-2626-49a5-a7ef-25f0d96be762"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:25:42.887447 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.887330 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-service-ca" (OuterVolumeSpecName: "service-ca") pod "fe2034cd-2626-49a5-a7ef-25f0d96be762" (UID: "fe2034cd-2626-49a5-a7ef-25f0d96be762"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:25:42.887447 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.887436 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "fe2034cd-2626-49a5-a7ef-25f0d96be762" (UID: "fe2034cd-2626-49a5-a7ef-25f0d96be762"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:25:42.887659 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.887512 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "fe2034cd-2626-49a5-a7ef-25f0d96be762" (UID: "fe2034cd-2626-49a5-a7ef-25f0d96be762"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:25:42.889071 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.889051 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "fe2034cd-2626-49a5-a7ef-25f0d96be762" (UID: "fe2034cd-2626-49a5-a7ef-25f0d96be762"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:42.889146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.889070 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "fe2034cd-2626-49a5-a7ef-25f0d96be762" (UID: "fe2034cd-2626-49a5-a7ef-25f0d96be762"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:25:42.889146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.889118 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe2034cd-2626-49a5-a7ef-25f0d96be762-kube-api-access-t9stf" (OuterVolumeSpecName: "kube-api-access-t9stf") pod "fe2034cd-2626-49a5-a7ef-25f0d96be762" (UID: "fe2034cd-2626-49a5-a7ef-25f0d96be762"). InnerVolumeSpecName "kube-api-access-t9stf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:25:42.987978 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.987957 2569 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-oauth-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:42.987978 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.987976 2569 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:42.988103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.987987 2569 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-t9stf\" (UniqueName: \"kubernetes.io/projected/fe2034cd-2626-49a5-a7ef-25f0d96be762-kube-api-access-t9stf\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:42.988103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.987996 2569 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-service-ca\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:42.988103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.988005 2569 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe2034cd-2626-49a5-a7ef-25f0d96be762-console-serving-cert\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:42.988103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.988013 2569 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-trusted-ca-bundle\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:42.988103 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:42.988023 2569 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fe2034cd-2626-49a5-a7ef-25f0d96be762-oauth-serving-cert\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:25:43.600077 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.600054 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7f67678f9c-q8ssl_fe2034cd-2626-49a5-a7ef-25f0d96be762/console/0.log" Apr 17 09:25:43.600420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.600094 2569 generic.go:358] "Generic (PLEG): container finished" podID="fe2034cd-2626-49a5-a7ef-25f0d96be762" containerID="cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547" exitCode=2 Apr 17 09:25:43.600420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.600160 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7f67678f9c-q8ssl" Apr 17 09:25:43.600420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.600181 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7f67678f9c-q8ssl" event={"ID":"fe2034cd-2626-49a5-a7ef-25f0d96be762","Type":"ContainerDied","Data":"cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547"} Apr 17 09:25:43.600420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.600223 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7f67678f9c-q8ssl" event={"ID":"fe2034cd-2626-49a5-a7ef-25f0d96be762","Type":"ContainerDied","Data":"9880de5b36768a36b934d1dbe0aeafaaa2d1ff163fe06d09745b79b7f4739741"} Apr 17 09:25:43.600420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.600240 2569 scope.go:117] "RemoveContainer" containerID="cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547" Apr 17 09:25:43.609593 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.609577 2569 scope.go:117] "RemoveContainer" containerID="cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547" Apr 17 09:25:43.609879 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:25:43.609856 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547\": container with ID starting with cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547 not found: ID does not exist" containerID="cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547" Apr 17 09:25:43.609936 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.609889 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547"} err="failed to get container status \"cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547\": rpc error: code = NotFound desc = could not find container \"cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547\": container with ID starting with cd71aa84cc4d5f37e5399fd34d6522dbdcce47921e8cc77035566888bbfa7547 not found: ID does not exist" Apr 17 09:25:43.620461 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.620436 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7f67678f9c-q8ssl"] Apr 17 09:25:43.623224 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.623196 2569 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-7f67678f9c-q8ssl"] Apr 17 09:25:43.706627 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:25:43.706601 2569 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe2034cd-2626-49a5-a7ef-25f0d96be762" path="/var/lib/kubelet/pods/fe2034cd-2626-49a5-a7ef-25f0d96be762/volumes" Apr 17 09:26:07.585580 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:07.585554 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/2.log" Apr 17 09:26:07.586043 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:07.586017 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/2.log" Apr 17 09:26:07.588899 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:07.588882 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/0.log" Apr 17 09:26:07.589200 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:07.589183 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/0.log" Apr 17 09:26:07.594708 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:07.594690 2569 kubelet.go:1628] "Image garbage collection succeeded" Apr 17 09:26:28.889951 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:28.889918 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-76d7958b94-xcmhv"] Apr 17 09:26:28.890458 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:28.890258 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fe2034cd-2626-49a5-a7ef-25f0d96be762" containerName="console" Apr 17 09:26:28.890458 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:28.890270 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe2034cd-2626-49a5-a7ef-25f0d96be762" containerName="console" Apr 17 09:26:28.890458 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:28.890327 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="fe2034cd-2626-49a5-a7ef-25f0d96be762" containerName="console" Apr 17 09:26:28.893237 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:28.893216 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:28.902960 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:28.902937 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76d7958b94-xcmhv"] Apr 17 09:26:29.019022 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.018986 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cb856f42-7746-4730-8167-6e4878870ba1-console-oauth-config\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.019186 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.019028 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-trusted-ca-bundle\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.019186 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.019065 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfscf\" (UniqueName: \"kubernetes.io/projected/cb856f42-7746-4730-8167-6e4878870ba1-kube-api-access-tfscf\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.019186 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.019149 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-oauth-serving-cert\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.019313 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.019185 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-console-config\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.019313 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.019218 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-service-ca\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.019313 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.019259 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb856f42-7746-4730-8167-6e4878870ba1-console-serving-cert\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.119637 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.119612 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-oauth-serving-cert\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.119753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.119645 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-console-config\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.119753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.119675 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-service-ca\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.119753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.119705 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb856f42-7746-4730-8167-6e4878870ba1-console-serving-cert\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.119753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.119747 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cb856f42-7746-4730-8167-6e4878870ba1-console-oauth-config\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.119965 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.119774 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-trusted-ca-bundle\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.119965 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.119800 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tfscf\" (UniqueName: \"kubernetes.io/projected/cb856f42-7746-4730-8167-6e4878870ba1-kube-api-access-tfscf\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.120400 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.120377 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-oauth-serving-cert\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.120566 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.120390 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-service-ca\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.120669 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.120470 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-console-config\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.120669 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.120583 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb856f42-7746-4730-8167-6e4878870ba1-trusted-ca-bundle\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.122047 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.122027 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb856f42-7746-4730-8167-6e4878870ba1-console-serving-cert\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.122127 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.122088 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cb856f42-7746-4730-8167-6e4878870ba1-console-oauth-config\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.127398 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.127379 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfscf\" (UniqueName: \"kubernetes.io/projected/cb856f42-7746-4730-8167-6e4878870ba1-kube-api-access-tfscf\") pod \"console-76d7958b94-xcmhv\" (UID: \"cb856f42-7746-4730-8167-6e4878870ba1\") " pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.204311 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.204246 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:29.520751 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.520730 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76d7958b94-xcmhv"] Apr 17 09:26:29.523157 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:26:29.523127 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb856f42_7746_4730_8167_6e4878870ba1.slice/crio-7bd5a0abab3490a990a45b7a6e20b9fd60de86b286a4c9de81b95533eddd3c0c WatchSource:0}: Error finding container 7bd5a0abab3490a990a45b7a6e20b9fd60de86b286a4c9de81b95533eddd3c0c: Status 404 returned error can't find the container with id 7bd5a0abab3490a990a45b7a6e20b9fd60de86b286a4c9de81b95533eddd3c0c Apr 17 09:26:29.524903 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.524886 2569 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 09:26:29.733389 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.733353 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76d7958b94-xcmhv" event={"ID":"cb856f42-7746-4730-8167-6e4878870ba1","Type":"ContainerStarted","Data":"551f70d6ecbe933a7113897cc90e52b4700bddacecf5331d6f14602086a2a378"} Apr 17 09:26:29.733389 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.733391 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76d7958b94-xcmhv" event={"ID":"cb856f42-7746-4730-8167-6e4878870ba1","Type":"ContainerStarted","Data":"7bd5a0abab3490a990a45b7a6e20b9fd60de86b286a4c9de81b95533eddd3c0c"} Apr 17 09:26:29.750873 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:29.750829 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-76d7958b94-xcmhv" podStartSLOduration=1.750799999 podStartE2EDuration="1.750799999s" podCreationTimestamp="2026-04-17 09:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:26:29.748379742 +0000 UTC m=+322.589416951" watchObservedRunningTime="2026-04-17 09:26:29.750799999 +0000 UTC m=+322.591837209" Apr 17 09:26:39.204790 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:39.204754 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:39.205222 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:39.204802 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:39.209461 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:39.209442 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:39.763482 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:39.763455 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-76d7958b94-xcmhv" Apr 17 09:26:39.809892 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:39.809863 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-fbd67bd86-86j2d"] Apr 17 09:26:52.119296 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.119229 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-k6vh8"] Apr 17 09:26:52.122442 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.122426 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.124781 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.124762 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 17 09:26:52.131149 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.131125 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-k6vh8"] Apr 17 09:26:52.191644 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.191614 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f495a154-8fa1-4339-ac70-804a4664bc14-kubelet-config\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.191760 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.191649 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f495a154-8fa1-4339-ac70-804a4664bc14-original-pull-secret\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.191760 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.191672 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f495a154-8fa1-4339-ac70-804a4664bc14-dbus\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.292564 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.292535 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f495a154-8fa1-4339-ac70-804a4664bc14-kubelet-config\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.292564 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.292567 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f495a154-8fa1-4339-ac70-804a4664bc14-original-pull-secret\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.292736 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.292585 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f495a154-8fa1-4339-ac70-804a4664bc14-dbus\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.292736 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.292660 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f495a154-8fa1-4339-ac70-804a4664bc14-kubelet-config\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.292736 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.292722 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f495a154-8fa1-4339-ac70-804a4664bc14-dbus\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.294639 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.294625 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f495a154-8fa1-4339-ac70-804a4664bc14-original-pull-secret\") pod \"global-pull-secret-syncer-k6vh8\" (UID: \"f495a154-8fa1-4339-ac70-804a4664bc14\") " pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.431503 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.431453 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-k6vh8" Apr 17 09:26:52.540764 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.540741 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-k6vh8"] Apr 17 09:26:52.542473 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:26:52.542440 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf495a154_8fa1_4339_ac70_804a4664bc14.slice/crio-aefebf08d821222352ef548987d80eeea5b7c035e552699a9c7498259a7cf62e WatchSource:0}: Error finding container aefebf08d821222352ef548987d80eeea5b7c035e552699a9c7498259a7cf62e: Status 404 returned error can't find the container with id aefebf08d821222352ef548987d80eeea5b7c035e552699a9c7498259a7cf62e Apr 17 09:26:52.805529 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:52.805499 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-k6vh8" event={"ID":"f495a154-8fa1-4339-ac70-804a4664bc14","Type":"ContainerStarted","Data":"aefebf08d821222352ef548987d80eeea5b7c035e552699a9c7498259a7cf62e"} Apr 17 09:26:56.819446 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:56.819414 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-k6vh8" event={"ID":"f495a154-8fa1-4339-ac70-804a4664bc14","Type":"ContainerStarted","Data":"a34aeb5983214a66373a0a3da6104e19c362cc73f1632f3bdf4305f6a9105390"} Apr 17 09:26:56.835903 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:26:56.835857 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-k6vh8" podStartSLOduration=1.080232342 podStartE2EDuration="4.835800413s" podCreationTimestamp="2026-04-17 09:26:52 +0000 UTC" firstStartedPulling="2026-04-17 09:26:52.544080577 +0000 UTC m=+345.385117766" lastFinishedPulling="2026-04-17 09:26:56.299648644 +0000 UTC m=+349.140685837" observedRunningTime="2026-04-17 09:26:56.834468623 +0000 UTC m=+349.675505845" watchObservedRunningTime="2026-04-17 09:26:56.835800413 +0000 UTC m=+349.676837620" Apr 17 09:27:04.830128 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:04.830081 2569 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-fbd67bd86-86j2d" podUID="dacde601-d4d3-4635-bf77-62a3cd0c1499" containerName="console" containerID="cri-o://ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d" gracePeriod=15 Apr 17 09:27:05.065652 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.065631 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-fbd67bd86-86j2d_dacde601-d4d3-4635-bf77-62a3cd0c1499/console/0.log" Apr 17 09:27:05.065762 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.065690 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:27:05.188757 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.188673 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-oauth-config\") pod \"dacde601-d4d3-4635-bf77-62a3cd0c1499\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " Apr 17 09:27:05.188757 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.188718 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-trusted-ca-bundle\") pod \"dacde601-d4d3-4635-bf77-62a3cd0c1499\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " Apr 17 09:27:05.188757 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.188743 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-service-ca\") pod \"dacde601-d4d3-4635-bf77-62a3cd0c1499\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " Apr 17 09:27:05.189051 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.188774 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-oauth-serving-cert\") pod \"dacde601-d4d3-4635-bf77-62a3cd0c1499\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " Apr 17 09:27:05.189051 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.188797 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99c8c\" (UniqueName: \"kubernetes.io/projected/dacde601-d4d3-4635-bf77-62a3cd0c1499-kube-api-access-99c8c\") pod \"dacde601-d4d3-4635-bf77-62a3cd0c1499\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " Apr 17 09:27:05.189051 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.188880 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-serving-cert\") pod \"dacde601-d4d3-4635-bf77-62a3cd0c1499\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " Apr 17 09:27:05.189051 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.188924 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-config\") pod \"dacde601-d4d3-4635-bf77-62a3cd0c1499\" (UID: \"dacde601-d4d3-4635-bf77-62a3cd0c1499\") " Apr 17 09:27:05.189249 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.189160 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "dacde601-d4d3-4635-bf77-62a3cd0c1499" (UID: "dacde601-d4d3-4635-bf77-62a3cd0c1499"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:27:05.189249 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.189201 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-service-ca" (OuterVolumeSpecName: "service-ca") pod "dacde601-d4d3-4635-bf77-62a3cd0c1499" (UID: "dacde601-d4d3-4635-bf77-62a3cd0c1499"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:27:05.189249 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.189228 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "dacde601-d4d3-4635-bf77-62a3cd0c1499" (UID: "dacde601-d4d3-4635-bf77-62a3cd0c1499"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:27:05.189456 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.189431 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-config" (OuterVolumeSpecName: "console-config") pod "dacde601-d4d3-4635-bf77-62a3cd0c1499" (UID: "dacde601-d4d3-4635-bf77-62a3cd0c1499"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:27:05.191099 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.191067 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "dacde601-d4d3-4635-bf77-62a3cd0c1499" (UID: "dacde601-d4d3-4635-bf77-62a3cd0c1499"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:27:05.191419 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.191400 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "dacde601-d4d3-4635-bf77-62a3cd0c1499" (UID: "dacde601-d4d3-4635-bf77-62a3cd0c1499"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:27:05.191471 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.191437 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dacde601-d4d3-4635-bf77-62a3cd0c1499-kube-api-access-99c8c" (OuterVolumeSpecName: "kube-api-access-99c8c") pod "dacde601-d4d3-4635-bf77-62a3cd0c1499" (UID: "dacde601-d4d3-4635-bf77-62a3cd0c1499"). InnerVolumeSpecName "kube-api-access-99c8c". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:27:05.289760 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.289742 2569 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-oauth-serving-cert\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:05.289760 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.289760 2569 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-99c8c\" (UniqueName: \"kubernetes.io/projected/dacde601-d4d3-4635-bf77-62a3cd0c1499-kube-api-access-99c8c\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:05.289899 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.289770 2569 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-serving-cert\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:05.289899 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.289780 2569 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:05.289899 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.289788 2569 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dacde601-d4d3-4635-bf77-62a3cd0c1499-console-oauth-config\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:05.289899 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.289796 2569 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-trusted-ca-bundle\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:05.289899 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.289805 2569 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dacde601-d4d3-4635-bf77-62a3cd0c1499-service-ca\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:05.848347 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.848326 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-fbd67bd86-86j2d_dacde601-d4d3-4635-bf77-62a3cd0c1499/console/0.log" Apr 17 09:27:05.848722 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.848364 2569 generic.go:358] "Generic (PLEG): container finished" podID="dacde601-d4d3-4635-bf77-62a3cd0c1499" containerID="ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d" exitCode=2 Apr 17 09:27:05.848722 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.848395 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-fbd67bd86-86j2d" event={"ID":"dacde601-d4d3-4635-bf77-62a3cd0c1499","Type":"ContainerDied","Data":"ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d"} Apr 17 09:27:05.848722 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.848424 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-fbd67bd86-86j2d" Apr 17 09:27:05.848722 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.848433 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-fbd67bd86-86j2d" event={"ID":"dacde601-d4d3-4635-bf77-62a3cd0c1499","Type":"ContainerDied","Data":"4333d05a468d83d6d393554438c56430b20e85046efcce1ba49e39f5fcebaee4"} Apr 17 09:27:05.848722 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.848448 2569 scope.go:117] "RemoveContainer" containerID="ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d" Apr 17 09:27:05.859097 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.859074 2569 scope.go:117] "RemoveContainer" containerID="ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d" Apr 17 09:27:05.859378 ip-10-0-137-152 kubenswrapper[2569]: E0417 09:27:05.859359 2569 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d\": container with ID starting with ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d not found: ID does not exist" containerID="ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d" Apr 17 09:27:05.859449 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.859388 2569 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d"} err="failed to get container status \"ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d\": rpc error: code = NotFound desc = could not find container \"ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d\": container with ID starting with ea9bdbafe42d063fc5da7de56975c1c5f7680de56628fc8f95f96f9bc560238d not found: ID does not exist" Apr 17 09:27:05.866369 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.866346 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-fbd67bd86-86j2d"] Apr 17 09:27:05.869943 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:05.869924 2569 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-fbd67bd86-86j2d"] Apr 17 09:27:07.706630 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:07.706601 2569 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dacde601-d4d3-4635-bf77-62a3cd0c1499" path="/var/lib/kubelet/pods/dacde601-d4d3-4635-bf77-62a3cd0c1499/volumes" Apr 17 09:27:17.860528 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.860494 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k"] Apr 17 09:27:17.861041 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.860966 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dacde601-d4d3-4635-bf77-62a3cd0c1499" containerName="console" Apr 17 09:27:17.861041 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.860983 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="dacde601-d4d3-4635-bf77-62a3cd0c1499" containerName="console" Apr 17 09:27:17.861162 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.861086 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="dacde601-d4d3-4635-bf77-62a3cd0c1499" containerName="console" Apr 17 09:27:17.864047 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.864026 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:17.866456 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.866437 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 17 09:27:17.866559 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.866440 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-xpt74\"" Apr 17 09:27:17.866559 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.866440 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 17 09:27:17.872113 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.872089 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k"] Apr 17 09:27:17.983151 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.983100 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lwkm\" (UniqueName: \"kubernetes.io/projected/522a0786-5996-4d11-9107-7aa15f8e5738-kube-api-access-6lwkm\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:17.983151 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.983156 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:17.983312 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:17.983220 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.083549 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.083522 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6lwkm\" (UniqueName: \"kubernetes.io/projected/522a0786-5996-4d11-9107-7aa15f8e5738-kube-api-access-6lwkm\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.083644 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.083556 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.083644 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.083581 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.084024 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.084005 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.084067 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.084021 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.091377 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.091356 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lwkm\" (UniqueName: \"kubernetes.io/projected/522a0786-5996-4d11-9107-7aa15f8e5738-kube-api-access-6lwkm\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.173266 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.173223 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:18.284201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.284176 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k"] Apr 17 09:27:18.286064 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:27:18.286037 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod522a0786_5996_4d11_9107_7aa15f8e5738.slice/crio-d998239241933fe55f12a80ee4b57adeeebfb096ec05e51fd8cfc2e50019d9f3 WatchSource:0}: Error finding container d998239241933fe55f12a80ee4b57adeeebfb096ec05e51fd8cfc2e50019d9f3: Status 404 returned error can't find the container with id d998239241933fe55f12a80ee4b57adeeebfb096ec05e51fd8cfc2e50019d9f3 Apr 17 09:27:18.887732 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:18.887692 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" event={"ID":"522a0786-5996-4d11-9107-7aa15f8e5738","Type":"ContainerStarted","Data":"d998239241933fe55f12a80ee4b57adeeebfb096ec05e51fd8cfc2e50019d9f3"} Apr 17 09:27:23.905724 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:23.905684 2569 generic.go:358] "Generic (PLEG): container finished" podID="522a0786-5996-4d11-9107-7aa15f8e5738" containerID="d2a38ad115deb111013ee3e6073f301f778762acfafdfd9ac98cfc5a54ea4c8d" exitCode=0 Apr 17 09:27:23.906165 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:23.905772 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" event={"ID":"522a0786-5996-4d11-9107-7aa15f8e5738","Type":"ContainerDied","Data":"d2a38ad115deb111013ee3e6073f301f778762acfafdfd9ac98cfc5a54ea4c8d"} Apr 17 09:27:25.913198 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:25.913138 2569 generic.go:358] "Generic (PLEG): container finished" podID="522a0786-5996-4d11-9107-7aa15f8e5738" containerID="7b2d8be58a9bfa3467175a245649613854b8a8ca85fefc9626808a9df76e323a" exitCode=0 Apr 17 09:27:25.913487 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:25.913223 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" event={"ID":"522a0786-5996-4d11-9107-7aa15f8e5738","Type":"ContainerDied","Data":"7b2d8be58a9bfa3467175a245649613854b8a8ca85fefc9626808a9df76e323a"} Apr 17 09:27:31.933412 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:31.933384 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" event={"ID":"522a0786-5996-4d11-9107-7aa15f8e5738","Type":"ContainerStarted","Data":"412030bea58c56716a7c1ab77dba5c78af339364a280a6a2969b79e11e7b9f86"} Apr 17 09:27:31.949408 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:31.949359 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" podStartSLOduration=1.412222115 podStartE2EDuration="14.94934671s" podCreationTimestamp="2026-04-17 09:27:17 +0000 UTC" firstStartedPulling="2026-04-17 09:27:18.290514955 +0000 UTC m=+371.131552157" lastFinishedPulling="2026-04-17 09:27:31.827639548 +0000 UTC m=+384.668676752" observedRunningTime="2026-04-17 09:27:31.948103473 +0000 UTC m=+384.789140680" watchObservedRunningTime="2026-04-17 09:27:31.94934671 +0000 UTC m=+384.790383921" Apr 17 09:27:32.937628 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:32.937594 2569 generic.go:358] "Generic (PLEG): container finished" podID="522a0786-5996-4d11-9107-7aa15f8e5738" containerID="412030bea58c56716a7c1ab77dba5c78af339364a280a6a2969b79e11e7b9f86" exitCode=0 Apr 17 09:27:32.938065 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:32.937713 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" event={"ID":"522a0786-5996-4d11-9107-7aa15f8e5738","Type":"ContainerDied","Data":"412030bea58c56716a7c1ab77dba5c78af339364a280a6a2969b79e11e7b9f86"} Apr 17 09:27:34.057783 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.057761 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:34.226002 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.225930 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-bundle\") pod \"522a0786-5996-4d11-9107-7aa15f8e5738\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " Apr 17 09:27:34.226002 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.225963 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-util\") pod \"522a0786-5996-4d11-9107-7aa15f8e5738\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " Apr 17 09:27:34.226216 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.226033 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lwkm\" (UniqueName: \"kubernetes.io/projected/522a0786-5996-4d11-9107-7aa15f8e5738-kube-api-access-6lwkm\") pod \"522a0786-5996-4d11-9107-7aa15f8e5738\" (UID: \"522a0786-5996-4d11-9107-7aa15f8e5738\") " Apr 17 09:27:34.226437 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.226408 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-bundle" (OuterVolumeSpecName: "bundle") pod "522a0786-5996-4d11-9107-7aa15f8e5738" (UID: "522a0786-5996-4d11-9107-7aa15f8e5738"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:27:34.228042 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.228019 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522a0786-5996-4d11-9107-7aa15f8e5738-kube-api-access-6lwkm" (OuterVolumeSpecName: "kube-api-access-6lwkm") pod "522a0786-5996-4d11-9107-7aa15f8e5738" (UID: "522a0786-5996-4d11-9107-7aa15f8e5738"). InnerVolumeSpecName "kube-api-access-6lwkm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:27:34.230644 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.230619 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-util" (OuterVolumeSpecName: "util") pod "522a0786-5996-4d11-9107-7aa15f8e5738" (UID: "522a0786-5996-4d11-9107-7aa15f8e5738"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:27:34.327552 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.327533 2569 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6lwkm\" (UniqueName: \"kubernetes.io/projected/522a0786-5996-4d11-9107-7aa15f8e5738-kube-api-access-6lwkm\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:34.327552 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.327554 2569 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-bundle\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:34.327664 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.327563 2569 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/522a0786-5996-4d11-9107-7aa15f8e5738-util\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:27:34.944894 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.944859 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" event={"ID":"522a0786-5996-4d11-9107-7aa15f8e5738","Type":"ContainerDied","Data":"d998239241933fe55f12a80ee4b57adeeebfb096ec05e51fd8cfc2e50019d9f3"} Apr 17 09:27:34.944894 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.944893 2569 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d998239241933fe55f12a80ee4b57adeeebfb096ec05e51fd8cfc2e50019d9f3" Apr 17 09:27:34.945102 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:34.944987 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e524l9k" Apr 17 09:27:40.155649 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.155618 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh"] Apr 17 09:27:40.156025 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.155934 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="522a0786-5996-4d11-9107-7aa15f8e5738" containerName="extract" Apr 17 09:27:40.156025 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.155945 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="522a0786-5996-4d11-9107-7aa15f8e5738" containerName="extract" Apr 17 09:27:40.156025 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.155957 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="522a0786-5996-4d11-9107-7aa15f8e5738" containerName="util" Apr 17 09:27:40.156025 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.155962 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="522a0786-5996-4d11-9107-7aa15f8e5738" containerName="util" Apr 17 09:27:40.156025 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.155975 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="522a0786-5996-4d11-9107-7aa15f8e5738" containerName="pull" Apr 17 09:27:40.156025 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.155980 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="522a0786-5996-4d11-9107-7aa15f8e5738" containerName="pull" Apr 17 09:27:40.156221 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.156067 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="522a0786-5996-4d11-9107-7aa15f8e5738" containerName="extract" Apr 17 09:27:40.160197 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.160180 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.162473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.162448 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager-operator\"/\"cert-manager-operator-controller-manager-dockercfg-x5cr2\"" Apr 17 09:27:40.162473 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.162467 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:27:40.162655 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.162467 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"kube-root-ca.crt\"" Apr 17 09:27:40.168990 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.168968 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh"] Apr 17 09:27:40.272673 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.272649 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/623733f7-f3c3-46fe-bf73-f7f6e6744dfd-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pszfh\" (UID: \"623733f7-f3c3-46fe-bf73-f7f6e6744dfd\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.272802 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.272704 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r9dk\" (UniqueName: \"kubernetes.io/projected/623733f7-f3c3-46fe-bf73-f7f6e6744dfd-kube-api-access-5r9dk\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pszfh\" (UID: \"623733f7-f3c3-46fe-bf73-f7f6e6744dfd\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.373654 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.373619 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/623733f7-f3c3-46fe-bf73-f7f6e6744dfd-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pszfh\" (UID: \"623733f7-f3c3-46fe-bf73-f7f6e6744dfd\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.373768 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.373692 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5r9dk\" (UniqueName: \"kubernetes.io/projected/623733f7-f3c3-46fe-bf73-f7f6e6744dfd-kube-api-access-5r9dk\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pszfh\" (UID: \"623733f7-f3c3-46fe-bf73-f7f6e6744dfd\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.374085 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.374068 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/623733f7-f3c3-46fe-bf73-f7f6e6744dfd-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pszfh\" (UID: \"623733f7-f3c3-46fe-bf73-f7f6e6744dfd\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.381895 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.381873 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r9dk\" (UniqueName: \"kubernetes.io/projected/623733f7-f3c3-46fe-bf73-f7f6e6744dfd-kube-api-access-5r9dk\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pszfh\" (UID: \"623733f7-f3c3-46fe-bf73-f7f6e6744dfd\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.469200 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.469149 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" Apr 17 09:27:40.587560 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.587534 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh"] Apr 17 09:27:40.589794 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:27:40.589765 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod623733f7_f3c3_46fe_bf73_f7f6e6744dfd.slice/crio-123506e1003f9a58bbfedbab6001e01e52385bb52616176a29a3589e28e0742c WatchSource:0}: Error finding container 123506e1003f9a58bbfedbab6001e01e52385bb52616176a29a3589e28e0742c: Status 404 returned error can't find the container with id 123506e1003f9a58bbfedbab6001e01e52385bb52616176a29a3589e28e0742c Apr 17 09:27:40.964043 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:40.964007 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" event={"ID":"623733f7-f3c3-46fe-bf73-f7f6e6744dfd","Type":"ContainerStarted","Data":"123506e1003f9a58bbfedbab6001e01e52385bb52616176a29a3589e28e0742c"} Apr 17 09:27:42.971347 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:42.971316 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" event={"ID":"623733f7-f3c3-46fe-bf73-f7f6e6744dfd","Type":"ContainerStarted","Data":"adfc38dbe55b575dde938231e189f55248943e99de1759a5a13162845125739c"} Apr 17 09:27:42.993106 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:42.993037 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pszfh" podStartSLOduration=1.268964529 podStartE2EDuration="2.993021261s" podCreationTimestamp="2026-04-17 09:27:40 +0000 UTC" firstStartedPulling="2026-04-17 09:27:40.592626603 +0000 UTC m=+393.433663792" lastFinishedPulling="2026-04-17 09:27:42.316683326 +0000 UTC m=+395.157720524" observedRunningTime="2026-04-17 09:27:42.991962477 +0000 UTC m=+395.832999712" watchObservedRunningTime="2026-04-17 09:27:42.993021261 +0000 UTC m=+395.834058479" Apr 17 09:27:48.438512 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.438470 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-mwtzj"] Apr 17 09:27:48.441901 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.441864 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.444063 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.444041 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Apr 17 09:27:48.444284 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.444268 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Apr 17 09:27:48.444852 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.444837 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-gmjkv\"" Apr 17 09:27:48.448369 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.448345 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-mwtzj"] Apr 17 09:27:48.537408 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.537381 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v7ns\" (UniqueName: \"kubernetes.io/projected/d3c319c2-c3de-4707-9c21-a343b62fa383-kube-api-access-5v7ns\") pod \"cert-manager-webhook-597b96b99b-mwtzj\" (UID: \"d3c319c2-c3de-4707-9c21-a343b62fa383\") " pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.537508 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.537437 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d3c319c2-c3de-4707-9c21-a343b62fa383-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-mwtzj\" (UID: \"d3c319c2-c3de-4707-9c21-a343b62fa383\") " pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.637752 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.637728 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d3c319c2-c3de-4707-9c21-a343b62fa383-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-mwtzj\" (UID: \"d3c319c2-c3de-4707-9c21-a343b62fa383\") " pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.637868 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.637775 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5v7ns\" (UniqueName: \"kubernetes.io/projected/d3c319c2-c3de-4707-9c21-a343b62fa383-kube-api-access-5v7ns\") pod \"cert-manager-webhook-597b96b99b-mwtzj\" (UID: \"d3c319c2-c3de-4707-9c21-a343b62fa383\") " pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.644920 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.644899 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v7ns\" (UniqueName: \"kubernetes.io/projected/d3c319c2-c3de-4707-9c21-a343b62fa383-kube-api-access-5v7ns\") pod \"cert-manager-webhook-597b96b99b-mwtzj\" (UID: \"d3c319c2-c3de-4707-9c21-a343b62fa383\") " pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.645442 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.645425 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d3c319c2-c3de-4707-9c21-a343b62fa383-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-mwtzj\" (UID: \"d3c319c2-c3de-4707-9c21-a343b62fa383\") " pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.764561 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.764529 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:48.874172 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.874105 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-mwtzj"] Apr 17 09:27:48.876302 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:27:48.876276 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3c319c2_c3de_4707_9c21_a343b62fa383.slice/crio-a062a353e2b7b880c4064cbb23f7411742c97cf295e20f05c9d458edc4d781f0 WatchSource:0}: Error finding container a062a353e2b7b880c4064cbb23f7411742c97cf295e20f05c9d458edc4d781f0: Status 404 returned error can't find the container with id a062a353e2b7b880c4064cbb23f7411742c97cf295e20f05c9d458edc4d781f0 Apr 17 09:27:48.991041 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:48.991011 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" event={"ID":"d3c319c2-c3de-4707-9c21-a343b62fa383","Type":"ContainerStarted","Data":"a062a353e2b7b880c4064cbb23f7411742c97cf295e20f05c9d458edc4d781f0"} Apr 17 09:27:53.006420 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:53.006382 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" event={"ID":"d3c319c2-c3de-4707-9c21-a343b62fa383","Type":"ContainerStarted","Data":"63b046afffb54d3da1ff3fc4f7b664c4cfa1b045f7ad533820e8b102d6c09096"} Apr 17 09:27:53.006803 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:53.006449 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:27:53.021179 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:53.021128 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" podStartSLOduration=1.838379152 podStartE2EDuration="5.021114704s" podCreationTimestamp="2026-04-17 09:27:48 +0000 UTC" firstStartedPulling="2026-04-17 09:27:48.881522105 +0000 UTC m=+401.722559294" lastFinishedPulling="2026-04-17 09:27:52.064257647 +0000 UTC m=+404.905294846" observedRunningTime="2026-04-17 09:27:53.01998578 +0000 UTC m=+405.861022989" watchObservedRunningTime="2026-04-17 09:27:53.021114704 +0000 UTC m=+405.862151915" Apr 17 09:27:59.011086 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:27:59.011061 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-597b96b99b-mwtzj" Apr 17 09:28:04.752271 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.752243 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-759f64656b-ck8cb"] Apr 17 09:28:04.802064 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.802040 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-759f64656b-ck8cb"] Apr 17 09:28:04.802201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.802125 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:04.804298 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.804278 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-dockercfg-pzn75\"" Apr 17 09:28:04.852154 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.852134 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmf6c\" (UniqueName: \"kubernetes.io/projected/cb97f620-b424-440a-8b87-c2fcaca373b2-kube-api-access-dmf6c\") pod \"cert-manager-759f64656b-ck8cb\" (UID: \"cb97f620-b424-440a-8b87-c2fcaca373b2\") " pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:04.852259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.852161 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb97f620-b424-440a-8b87-c2fcaca373b2-bound-sa-token\") pod \"cert-manager-759f64656b-ck8cb\" (UID: \"cb97f620-b424-440a-8b87-c2fcaca373b2\") " pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:04.952796 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.952770 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dmf6c\" (UniqueName: \"kubernetes.io/projected/cb97f620-b424-440a-8b87-c2fcaca373b2-kube-api-access-dmf6c\") pod \"cert-manager-759f64656b-ck8cb\" (UID: \"cb97f620-b424-440a-8b87-c2fcaca373b2\") " pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:04.952941 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.952800 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb97f620-b424-440a-8b87-c2fcaca373b2-bound-sa-token\") pod \"cert-manager-759f64656b-ck8cb\" (UID: \"cb97f620-b424-440a-8b87-c2fcaca373b2\") " pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:04.960396 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.960373 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb97f620-b424-440a-8b87-c2fcaca373b2-bound-sa-token\") pod \"cert-manager-759f64656b-ck8cb\" (UID: \"cb97f620-b424-440a-8b87-c2fcaca373b2\") " pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:04.960540 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:04.960522 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmf6c\" (UniqueName: \"kubernetes.io/projected/cb97f620-b424-440a-8b87-c2fcaca373b2-kube-api-access-dmf6c\") pod \"cert-manager-759f64656b-ck8cb\" (UID: \"cb97f620-b424-440a-8b87-c2fcaca373b2\") " pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:05.110695 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:05.110667 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-759f64656b-ck8cb" Apr 17 09:28:05.224201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:05.224169 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-759f64656b-ck8cb"] Apr 17 09:28:05.227396 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:28:05.227364 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb97f620_b424_440a_8b87_c2fcaca373b2.slice/crio-f372c7584fcfbda5a11787e1f545fe3053be04a0231de83efe8581d4f71ff4f0 WatchSource:0}: Error finding container f372c7584fcfbda5a11787e1f545fe3053be04a0231de83efe8581d4f71ff4f0: Status 404 returned error can't find the container with id f372c7584fcfbda5a11787e1f545fe3053be04a0231de83efe8581d4f71ff4f0 Apr 17 09:28:06.049391 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.049358 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-759f64656b-ck8cb" event={"ID":"cb97f620-b424-440a-8b87-c2fcaca373b2","Type":"ContainerStarted","Data":"aa0767608a986c4e5366479d00fda5ac1f3a6bd8f95fb61172dad7d7fe056010"} Apr 17 09:28:06.049772 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.049398 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-759f64656b-ck8cb" event={"ID":"cb97f620-b424-440a-8b87-c2fcaca373b2","Type":"ContainerStarted","Data":"f372c7584fcfbda5a11787e1f545fe3053be04a0231de83efe8581d4f71ff4f0"} Apr 17 09:28:06.068560 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.065225 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-759f64656b-ck8cb" podStartSLOduration=2.065209883 podStartE2EDuration="2.065209883s" podCreationTimestamp="2026-04-17 09:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:28:06.063302922 +0000 UTC m=+418.904340135" watchObservedRunningTime="2026-04-17 09:28:06.065209883 +0000 UTC m=+418.906247095" Apr 17 09:28:06.389010 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.388932 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj"] Apr 17 09:28:06.392753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.392737 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.395152 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.395126 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 17 09:28:06.395246 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.395131 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-xpt74\"" Apr 17 09:28:06.395246 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.395140 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 17 09:28:06.399098 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.399081 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj"] Apr 17 09:28:06.466647 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.466618 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrvz6\" (UniqueName: \"kubernetes.io/projected/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-kube-api-access-nrvz6\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.466751 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.466681 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-bundle\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.466751 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.466698 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-util\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.567707 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.567681 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nrvz6\" (UniqueName: \"kubernetes.io/projected/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-kube-api-access-nrvz6\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.567851 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.567761 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-bundle\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.567851 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.567789 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-util\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.568109 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.568089 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-bundle\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.568167 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.568116 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-util\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.575537 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.575521 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrvz6\" (UniqueName: \"kubernetes.io/projected/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-kube-api-access-nrvz6\") pod \"c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.702251 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.702178 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:06.817623 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:06.817600 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj"] Apr 17 09:28:06.820115 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:28:06.820074 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfad58b1_d8fc_45ae_b728_f3bd966d8b7b.slice/crio-d164bbea6aa74ef75fb17282fcc507a839f6da6282a66476a237cd14ed43a4a1 WatchSource:0}: Error finding container d164bbea6aa74ef75fb17282fcc507a839f6da6282a66476a237cd14ed43a4a1: Status 404 returned error can't find the container with id d164bbea6aa74ef75fb17282fcc507a839f6da6282a66476a237cd14ed43a4a1 Apr 17 09:28:07.053356 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:07.053325 2569 generic.go:358] "Generic (PLEG): container finished" podID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerID="d407bece0605aa71da4884c024fa785102a78a89b0654a3aaafa80d899e3dfb1" exitCode=0 Apr 17 09:28:07.053700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:07.053404 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" event={"ID":"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b","Type":"ContainerDied","Data":"d407bece0605aa71da4884c024fa785102a78a89b0654a3aaafa80d899e3dfb1"} Apr 17 09:28:07.053700 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:07.053430 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" event={"ID":"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b","Type":"ContainerStarted","Data":"d164bbea6aa74ef75fb17282fcc507a839f6da6282a66476a237cd14ed43a4a1"} Apr 17 09:28:10.070569 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:10.070538 2569 generic.go:358] "Generic (PLEG): container finished" podID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerID="08a24f99986bb017d162aa49647cd8b26df98362cf00bd1e5167681e5f04a769" exitCode=0 Apr 17 09:28:10.070933 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:10.070618 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" event={"ID":"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b","Type":"ContainerDied","Data":"08a24f99986bb017d162aa49647cd8b26df98362cf00bd1e5167681e5f04a769"} Apr 17 09:28:11.075383 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:11.075349 2569 generic.go:358] "Generic (PLEG): container finished" podID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerID="96ce65354fc8e88bce5a3bc26f923f42999d514044f300237956789822730efe" exitCode=0 Apr 17 09:28:11.075767 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:11.075387 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" event={"ID":"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b","Type":"ContainerDied","Data":"96ce65354fc8e88bce5a3bc26f923f42999d514044f300237956789822730efe"} Apr 17 09:28:12.196981 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.196958 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:12.313220 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.313196 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrvz6\" (UniqueName: \"kubernetes.io/projected/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-kube-api-access-nrvz6\") pod \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " Apr 17 09:28:12.313388 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.313267 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-bundle\") pod \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " Apr 17 09:28:12.313388 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.313294 2569 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-util\") pod \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\" (UID: \"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b\") " Apr 17 09:28:12.313631 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.313603 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-bundle" (OuterVolumeSpecName: "bundle") pod "dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" (UID: "dfad58b1-d8fc-45ae-b728-f3bd966d8b7b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:28:12.315249 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.315221 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-kube-api-access-nrvz6" (OuterVolumeSpecName: "kube-api-access-nrvz6") pod "dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" (UID: "dfad58b1-d8fc-45ae-b728-f3bd966d8b7b"). InnerVolumeSpecName "kube-api-access-nrvz6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:28:12.317380 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.317356 2569 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-util" (OuterVolumeSpecName: "util") pod "dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" (UID: "dfad58b1-d8fc-45ae-b728-f3bd966d8b7b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:28:12.414610 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.414561 2569 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nrvz6\" (UniqueName: \"kubernetes.io/projected/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-kube-api-access-nrvz6\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:28:12.414610 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.414581 2569 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-bundle\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:28:12.414610 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:12.414590 2569 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad58b1-d8fc-45ae-b728-f3bd966d8b7b-util\") on node \"ip-10-0-137-152.ec2.internal\" DevicePath \"\"" Apr 17 09:28:13.083965 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:13.083932 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" event={"ID":"dfad58b1-d8fc-45ae-b728-f3bd966d8b7b","Type":"ContainerDied","Data":"d164bbea6aa74ef75fb17282fcc507a839f6da6282a66476a237cd14ed43a4a1"} Apr 17 09:28:13.083965 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:13.083965 2569 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d164bbea6aa74ef75fb17282fcc507a839f6da6282a66476a237cd14ed43a4a1" Apr 17 09:28:13.084154 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:13.083962 2569 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c2ca89134faa49158137edb0141b62ea0c6a854657aff316cf72d9c78egb6fj" Apr 17 09:28:29.579857 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.579812 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w"] Apr 17 09:28:29.580215 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.580201 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerName="util" Apr 17 09:28:29.580259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.580217 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerName="util" Apr 17 09:28:29.580259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.580226 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerName="extract" Apr 17 09:28:29.580259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.580231 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerName="extract" Apr 17 09:28:29.580259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.580250 2569 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerName="pull" Apr 17 09:28:29.580259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.580255 2569 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerName="pull" Apr 17 09:28:29.580392 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.580305 2569 memory_manager.go:356] "RemoveStaleState removing state" podUID="dfad58b1-d8fc-45ae-b728-f3bd966d8b7b" containerName="extract" Apr 17 09:28:29.584515 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.584500 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.587665 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.587642 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-jobset-operator\"/\"webhook-server-cert\"" Apr 17 09:28:29.587857 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.587656 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-jobset-operator\"/\"jobset-manager-config\"" Apr 17 09:28:29.587986 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.587685 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-jobset-operator\"/\"metrics-server-cert\"" Apr 17 09:28:29.587986 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.587688 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-jobset-operator\"/\"jobset-controller-manager-dockercfg-bzr84\"" Apr 17 09:28:29.588140 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.587692 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-jobset-operator\"/\"kube-root-ca.crt\"" Apr 17 09:28:29.588140 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.587706 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-jobset-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:28:29.590640 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.590619 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w"] Apr 17 09:28:29.655699 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.655674 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b1f1b415-65b4-4815-bca2-becae1cbcecd-manager-config\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.655809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.655706 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjrf4\" (UniqueName: \"kubernetes.io/projected/b1f1b415-65b4-4815-bca2-becae1cbcecd-kube-api-access-kjrf4\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.655809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.655738 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1f1b415-65b4-4815-bca2-becae1cbcecd-metrics-certs\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.655809 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.655800 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b1f1b415-65b4-4815-bca2-becae1cbcecd-cert\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.756961 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.756930 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b1f1b415-65b4-4815-bca2-becae1cbcecd-manager-config\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.757110 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.756973 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kjrf4\" (UniqueName: \"kubernetes.io/projected/b1f1b415-65b4-4815-bca2-becae1cbcecd-kube-api-access-kjrf4\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.757110 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.757007 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1f1b415-65b4-4815-bca2-becae1cbcecd-metrics-certs\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.757110 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.757037 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b1f1b415-65b4-4815-bca2-becae1cbcecd-cert\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.757637 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.757617 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b1f1b415-65b4-4815-bca2-becae1cbcecd-manager-config\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.759259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.759242 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b1f1b415-65b4-4815-bca2-becae1cbcecd-cert\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.759400 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.759384 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1f1b415-65b4-4815-bca2-becae1cbcecd-metrics-certs\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.763630 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.763608 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjrf4\" (UniqueName: \"kubernetes.io/projected/b1f1b415-65b4-4815-bca2-becae1cbcecd-kube-api-access-kjrf4\") pod \"jobset-controller-manager-76864d57c7-kps8w\" (UID: \"b1f1b415-65b4-4815-bca2-becae1cbcecd\") " pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:29.895270 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:29.895212 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:30.008567 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:30.008534 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w"] Apr 17 09:28:30.010941 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:28:30.010911 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1f1b415_65b4_4815_bca2_becae1cbcecd.slice/crio-9ce3f5f52f39474c4d4b4613f96d6ea8c0e67f08e955815b083ebd2c0beec079 WatchSource:0}: Error finding container 9ce3f5f52f39474c4d4b4613f96d6ea8c0e67f08e955815b083ebd2c0beec079: Status 404 returned error can't find the container with id 9ce3f5f52f39474c4d4b4613f96d6ea8c0e67f08e955815b083ebd2c0beec079 Apr 17 09:28:30.140057 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:30.140027 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" event={"ID":"b1f1b415-65b4-4815-bca2-becae1cbcecd","Type":"ContainerStarted","Data":"9ce3f5f52f39474c4d4b4613f96d6ea8c0e67f08e955815b083ebd2c0beec079"} Apr 17 09:28:33.152648 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:33.152608 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" event={"ID":"b1f1b415-65b4-4815-bca2-becae1cbcecd","Type":"ContainerStarted","Data":"3909106461d33a465082bc533592ff879954db9f55698f9a795d7495e5f62946"} Apr 17 09:28:33.153025 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:33.152765 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:28:33.173885 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:33.169415 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" podStartSLOduration=1.547729058 podStartE2EDuration="4.169399037s" podCreationTimestamp="2026-04-17 09:28:29 +0000 UTC" firstStartedPulling="2026-04-17 09:28:30.012776675 +0000 UTC m=+442.853813865" lastFinishedPulling="2026-04-17 09:28:32.634446655 +0000 UTC m=+445.475483844" observedRunningTime="2026-04-17 09:28:33.167050666 +0000 UTC m=+446.008087877" watchObservedRunningTime="2026-04-17 09:28:33.169399037 +0000 UTC m=+446.010436248" Apr 17 09:28:44.161801 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:28:44.161773 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-jobset-operator/jobset-controller-manager-76864d57c7-kps8w" Apr 17 09:30:01.247666 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:01.247592 2569 ???:1] "http: TLS handshake error from 10.0.129.70:58506: EOF" Apr 17 09:30:01.249455 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:01.249436 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-k6vh8_f495a154-8fa1-4339-ac70-804a4664bc14/global-pull-secret-syncer/0.log" Apr 17 09:30:01.338328 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:01.338295 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-b65w4_d7b21dc1-e456-4f3a-827a-6d17899801ad/konnectivity-agent/0.log" Apr 17 09:30:01.423168 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:01.423137 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-137-152.ec2.internal_311677e33e183475ac3960f0885188ab/haproxy/0.log" Apr 17 09:30:04.651541 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.651519 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_ab8a248c-4463-4b8d-9ba6-1b2063937460/alertmanager/0.log" Apr 17 09:30:04.671351 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.671329 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_ab8a248c-4463-4b8d-9ba6-1b2063937460/config-reloader/0.log" Apr 17 09:30:04.690294 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.690271 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_ab8a248c-4463-4b8d-9ba6-1b2063937460/kube-rbac-proxy-web/0.log" Apr 17 09:30:04.710328 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.710303 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_ab8a248c-4463-4b8d-9ba6-1b2063937460/kube-rbac-proxy/0.log" Apr 17 09:30:04.730904 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.730874 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_ab8a248c-4463-4b8d-9ba6-1b2063937460/kube-rbac-proxy-metric/0.log" Apr 17 09:30:04.752549 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.752524 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_ab8a248c-4463-4b8d-9ba6-1b2063937460/prom-label-proxy/0.log" Apr 17 09:30:04.771153 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.771128 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_ab8a248c-4463-4b8d-9ba6-1b2063937460/init-config-reloader/0.log" Apr 17 09:30:04.802757 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.802727 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-9rpfg_22a71e37-6733-4843-a7c4-2027bbd017e5/cluster-monitoring-operator/0.log" Apr 17 09:30:04.891542 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.891513 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_metrics-server-544766d54-ktmxf_a45e7480-74e1-407a-85ac-03d53fb98a59/metrics-server/0.log" Apr 17 09:30:04.941839 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.941754 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-fm9gq_4222015d-39a6-4f11-b504-7a3dfc2576cf/node-exporter/0.log" Apr 17 09:30:04.961029 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.961002 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-fm9gq_4222015d-39a6-4f11-b504-7a3dfc2576cf/kube-rbac-proxy/0.log" Apr 17 09:30:04.978616 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:04.978591 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-fm9gq_4222015d-39a6-4f11-b504-7a3dfc2576cf/init-textfile/0.log" Apr 17 09:30:05.149857 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.149795 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-bttbq_54961548-4e6d-49bc-8671-8265a524aba5/kube-rbac-proxy-main/0.log" Apr 17 09:30:05.168467 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.168440 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-bttbq_54961548-4e6d-49bc-8671-8265a524aba5/kube-rbac-proxy-self/0.log" Apr 17 09:30:05.186465 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.186437 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-bttbq_54961548-4e6d-49bc-8671-8265a524aba5/openshift-state-metrics/0.log" Apr 17 09:30:05.446752 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.446727 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-admission-webhook-57cf98b594-7gvjw_0a8a1e02-ea5f-4784-9eaf-96a0ccf9068d/prometheus-operator-admission-webhook/0.log" Apr 17 09:30:05.496217 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.496191 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-9798999bf-dvwvs_bb3fcb85-ac31-47ee-b6e7-614532698db2/telemeter-client/0.log" Apr 17 09:30:05.525940 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.525917 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-9798999bf-dvwvs_bb3fcb85-ac31-47ee-b6e7-614532698db2/reload/0.log" Apr 17 09:30:05.544472 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.544443 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-9798999bf-dvwvs_bb3fcb85-ac31-47ee-b6e7-614532698db2/kube-rbac-proxy/0.log" Apr 17 09:30:05.575702 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.575658 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/thanos-query/0.log" Apr 17 09:30:05.598149 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.598123 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy-web/0.log" Apr 17 09:30:05.624188 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.624162 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy/0.log" Apr 17 09:30:05.644124 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.644103 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/prom-label-proxy/0.log" Apr 17 09:30:05.665759 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.665733 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy-rules/0.log" Apr 17 09:30:05.688753 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:05.688727 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-577f9447f9-8p9fd_52442855-76a8-4ca3-a454-53fe40b0e7d0/kube-rbac-proxy-metrics/0.log" Apr 17 09:30:06.961029 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:06.961000 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/2.log" Apr 17 09:30:06.965607 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:06.965590 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-z9nkj_ef4e2c76-47f7-42fb-a0e4-d95b609d0101/console-operator/3.log" Apr 17 09:30:07.262704 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.262682 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-76d7958b94-xcmhv_cb856f42-7746-4730-8167-6e4878870ba1/console/0.log" Apr 17 09:30:07.285791 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.285764 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-6bcc868b7-svf8b_a9543ac2-585d-407a-92ca-f7b7437d36f1/download-server/0.log" Apr 17 09:30:07.609109 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.609030 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r"] Apr 17 09:30:07.612374 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.612351 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.614470 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.614449 2569 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-xkrgt\"/\"default-dockercfg-v97xl\"" Apr 17 09:30:07.614577 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.614452 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-xkrgt\"/\"kube-root-ca.crt\"" Apr 17 09:30:07.615259 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.615245 2569 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-xkrgt\"/\"openshift-service-ca.crt\"" Apr 17 09:30:07.621361 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.621315 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r"] Apr 17 09:30:07.718418 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.718381 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-proc\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.718418 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.718432 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-podres\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.718615 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.718475 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-sys\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.718615 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.718500 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-lib-modules\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.718615 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.718519 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmxn\" (UniqueName: \"kubernetes.io/projected/412a72e9-0360-44a0-8bb5-057845ac58f7-kube-api-access-dcmxn\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819613 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819568 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-proc\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819768 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819639 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-podres\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819768 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819680 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-proc\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819768 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819726 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-sys\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819768 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819749 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-lib-modules\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819970 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819771 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmxn\" (UniqueName: \"kubernetes.io/projected/412a72e9-0360-44a0-8bb5-057845ac58f7-kube-api-access-dcmxn\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819970 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819786 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-podres\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819970 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819848 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-sys\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.819970 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.819886 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/412a72e9-0360-44a0-8bb5-057845ac58f7-lib-modules\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.826906 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.826889 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmxn\" (UniqueName: \"kubernetes.io/projected/412a72e9-0360-44a0-8bb5-057845ac58f7-kube-api-access-dcmxn\") pod \"perf-node-gather-daemonset-5mv8r\" (UID: \"412a72e9-0360-44a0-8bb5-057845ac58f7\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:07.923216 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:07.923134 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:08.042209 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.042183 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r"] Apr 17 09:30:08.044532 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:30:08.044501 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod412a72e9_0360_44a0_8bb5_057845ac58f7.slice/crio-f6194a60ded50bef205595e50f4a32b05075200d930f5c51d2729f229953b144 WatchSource:0}: Error finding container f6194a60ded50bef205595e50f4a32b05075200d930f5c51d2729f229953b144: Status 404 returned error can't find the container with id f6194a60ded50bef205595e50f4a32b05075200d930f5c51d2729f229953b144 Apr 17 09:30:08.261919 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.261892 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-27fm4_b171f472-b404-4138-bf91-727a857e3224/dns/0.log" Apr 17 09:30:08.280225 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.280206 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-27fm4_b171f472-b404-4138-bf91-727a857e3224/kube-rbac-proxy/0.log" Apr 17 09:30:08.403973 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.403950 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-b5dkm_9fcd330e-f463-420a-a91c-f2a6e23daebc/dns-node-resolver/0.log" Apr 17 09:30:08.479146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.479113 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" event={"ID":"412a72e9-0360-44a0-8bb5-057845ac58f7","Type":"ContainerStarted","Data":"5ab55de21c69fb4c028eea09e058db1b9730a803167a80fb2cd31ee372162a4f"} Apr 17 09:30:08.479146 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.479144 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" event={"ID":"412a72e9-0360-44a0-8bb5-057845ac58f7","Type":"ContainerStarted","Data":"f6194a60ded50bef205595e50f4a32b05075200d930f5c51d2729f229953b144"} Apr 17 09:30:08.479372 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.479215 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:08.495924 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.495882 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" podStartSLOduration=1.495869132 podStartE2EDuration="1.495869132s" podCreationTimestamp="2026-04-17 09:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:30:08.49460036 +0000 UTC m=+541.335637570" watchObservedRunningTime="2026-04-17 09:30:08.495869132 +0000 UTC m=+541.336906342" Apr 17 09:30:08.533101 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.533037 2569 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-799864bf59-2gpvr"] Apr 17 09:30:08.536230 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.536213 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.547830 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.547797 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-799864bf59-2gpvr"] Apr 17 09:30:08.624781 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.624758 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-service-ca\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.624921 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.624802 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-trusted-ca-bundle\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.624921 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.624844 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-config\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.624921 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.624899 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-oauth-config\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.625058 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.624952 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-serving-cert\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.625058 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.624980 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4qgw\" (UniqueName: \"kubernetes.io/projected/32607b6a-29d5-4ced-82e9-7ba0f3798456-kube-api-access-p4qgw\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.625058 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.625011 2569 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-oauth-serving-cert\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726116 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726085 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-oauth-config\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726309 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726130 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-serving-cert\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726309 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726150 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p4qgw\" (UniqueName: \"kubernetes.io/projected/32607b6a-29d5-4ced-82e9-7ba0f3798456-kube-api-access-p4qgw\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726309 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726269 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-oauth-serving-cert\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726476 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726339 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-service-ca\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726476 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726391 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-trusted-ca-bundle\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726476 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726420 2569 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-config\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.726987 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.726966 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-oauth-serving-cert\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.727108 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.727084 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-service-ca\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.727172 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.727112 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-config\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.727334 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.727316 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32607b6a-29d5-4ced-82e9-7ba0f3798456-trusted-ca-bundle\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.728616 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.728596 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-serving-cert\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.728687 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.728615 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/32607b6a-29d5-4ced-82e9-7ba0f3798456-console-oauth-config\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.733781 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.733764 2569 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4qgw\" (UniqueName: \"kubernetes.io/projected/32607b6a-29d5-4ced-82e9-7ba0f3798456-kube-api-access-p4qgw\") pod \"console-799864bf59-2gpvr\" (UID: \"32607b6a-29d5-4ced-82e9-7ba0f3798456\") " pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.846684 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.846609 2569 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:08.851680 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.851654 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_image-registry-fcffd4c79-f7kd9_5ab3de37-fccc-4ce9-9c52-a0d690b6311a/registry/0.log" Apr 17 09:30:08.912582 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.912557 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-t5ckx_09912815-97d7-47ef-939f-9f3815c992c1/node-ca/0.log" Apr 17 09:30:08.960207 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:08.960184 2569 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-799864bf59-2gpvr"] Apr 17 09:30:08.961848 ip-10-0-137-152 kubenswrapper[2569]: W0417 09:30:08.961793 2569 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32607b6a_29d5_4ced_82e9_7ba0f3798456.slice/crio-8194b1fb4e104220abfd0487214730e2a959eea7f408c58c815f909883c073b8 WatchSource:0}: Error finding container 8194b1fb4e104220abfd0487214730e2a959eea7f408c58c815f909883c073b8: Status 404 returned error can't find the container with id 8194b1fb4e104220abfd0487214730e2a959eea7f408c58c815f909883c073b8 Apr 17 09:30:09.483201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:09.483163 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-799864bf59-2gpvr" event={"ID":"32607b6a-29d5-4ced-82e9-7ba0f3798456","Type":"ContainerStarted","Data":"f72dc91acdade4613f38345ef921ebc0e7f6f218b05d3d2a12df0b7b8fbadf57"} Apr 17 09:30:09.483201 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:09.483204 2569 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-799864bf59-2gpvr" event={"ID":"32607b6a-29d5-4ced-82e9-7ba0f3798456","Type":"ContainerStarted","Data":"8194b1fb4e104220abfd0487214730e2a959eea7f408c58c815f909883c073b8"} Apr 17 09:30:09.499483 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:09.499440 2569 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-799864bf59-2gpvr" podStartSLOduration=1.499427004 podStartE2EDuration="1.499427004s" podCreationTimestamp="2026-04-17 09:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:30:09.49746233 +0000 UTC m=+542.338499542" watchObservedRunningTime="2026-04-17 09:30:09.499427004 +0000 UTC m=+542.340464214" Apr 17 09:30:09.908441 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:09.908414 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-pf795_e5c6f5f4-e252-4726-9992-25c01cae77a3/serve-healthcheck-canary/0.log" Apr 17 09:30:10.357997 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:10.357965 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-rfxfc_3d38afde-3d74-46a3-a5f7-e4cec8ee56d6/kube-rbac-proxy/0.log" Apr 17 09:30:10.375162 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:10.375131 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-rfxfc_3d38afde-3d74-46a3-a5f7-e4cec8ee56d6/exporter/0.log" Apr 17 09:30:10.393061 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:10.393039 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-rfxfc_3d38afde-3d74-46a3-a5f7-e4cec8ee56d6/extractor/0.log" Apr 17 09:30:11.829199 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:11.829173 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-jobset-operator_jobset-controller-manager-76864d57c7-kps8w_b1f1b415-65b4-4815-bca2-becae1cbcecd/manager/0.log" Apr 17 09:30:14.493218 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:14.493190 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-5mv8r" Apr 17 09:30:14.825021 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:14.824986 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-zj5rl_89378780-0d1b-40cd-930c-da0d27751523/kube-storage-version-migrator-operator/1.log" Apr 17 09:30:14.825842 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:14.825808 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-zj5rl_89378780-0d1b-40cd-930c-da0d27751523/kube-storage-version-migrator-operator/0.log" Apr 17 09:30:15.546974 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.546945 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9bwlv_e686455d-e53d-4a81-b3c4-74eb05f3da00/kube-multus/0.log" Apr 17 09:30:15.859487 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.859461 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-g7bkg_70fa4740-d7b3-4bc0-8829-d52ad2ac1144/kube-multus-additional-cni-plugins/0.log" Apr 17 09:30:15.877382 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.877357 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-g7bkg_70fa4740-d7b3-4bc0-8829-d52ad2ac1144/egress-router-binary-copy/0.log" Apr 17 09:30:15.894790 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.894772 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-g7bkg_70fa4740-d7b3-4bc0-8829-d52ad2ac1144/cni-plugins/0.log" Apr 17 09:30:15.912478 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.912447 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-g7bkg_70fa4740-d7b3-4bc0-8829-d52ad2ac1144/bond-cni-plugin/0.log" Apr 17 09:30:15.932338 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.932311 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-g7bkg_70fa4740-d7b3-4bc0-8829-d52ad2ac1144/routeoverride-cni/0.log" Apr 17 09:30:15.948282 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.948258 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-g7bkg_70fa4740-d7b3-4bc0-8829-d52ad2ac1144/whereabouts-cni-bincopy/0.log" Apr 17 09:30:15.968258 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:15.968235 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-g7bkg_70fa4740-d7b3-4bc0-8829-d52ad2ac1144/whereabouts-cni/0.log" Apr 17 09:30:16.092539 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:16.092459 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-4kbg5_833e0b85-0aba-4f60-bbf6-a215ed02831a/network-metrics-daemon/0.log" Apr 17 09:30:16.109756 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:16.109728 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-4kbg5_833e0b85-0aba-4f60-bbf6-a215ed02831a/kube-rbac-proxy/0.log" Apr 17 09:30:17.187762 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.187669 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-controller/0.log" Apr 17 09:30:17.203571 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.203542 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/0.log" Apr 17 09:30:17.205987 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.205965 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovn-acl-logging/1.log" Apr 17 09:30:17.221159 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.221134 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/kube-rbac-proxy-node/0.log" Apr 17 09:30:17.241294 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.241273 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/kube-rbac-proxy-ovn-metrics/0.log" Apr 17 09:30:17.257656 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.257638 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/northd/0.log" Apr 17 09:30:17.274584 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.274560 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/nbdb/0.log" Apr 17 09:30:17.294261 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.294241 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/sbdb/0.log" Apr 17 09:30:17.381930 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:17.381901 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd2rv_1ea35608-b7a5-4952-8b6c-f24a85b0ff89/ovnkube-controller/0.log" Apr 17 09:30:18.615309 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:18.615284 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-8894fc9bd-pzchj_abd1dcbf-ede8-4fa7-be3e-02a819a3421c/check-endpoints/0.log" Apr 17 09:30:18.635331 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:18.635306 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-2cw78_e338cd11-dfc1-43bb-8092-32e10f1e62a5/network-check-target-container/0.log" Apr 17 09:30:18.847418 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:18.847334 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:18.847418 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:18.847382 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:18.851782 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:18.851758 2569 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:19.439565 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:19.439535 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-dd7tv_2f0ca441-19bb-4396-adb5-c2b223622c08/iptables-alerter/0.log" Apr 17 09:30:19.518237 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:19.518213 2569 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-799864bf59-2gpvr" Apr 17 09:30:19.559963 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:19.559933 2569 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-76d7958b94-xcmhv"] Apr 17 09:30:20.043617 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:20.043586 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-rjm7r_0ed27ada-6c47-4d7b-9fe8-b0a6afea7e32/tuned/0.log" Apr 17 09:30:21.541652 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:21.541625 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-6dc5bdb6b4-pdqft_8f81e523-bc6a-40b3-8e85-891b04d3707b/cluster-samples-operator/0.log" Apr 17 09:30:21.555011 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:21.554986 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-6dc5bdb6b4-pdqft_8f81e523-bc6a-40b3-8e85-891b04d3707b/cluster-samples-operator-watch/0.log" Apr 17 09:30:22.983719 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:22.983645 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-8l4wz_904ead54-94c1-4e84-9851-0c8506ed7bbc/csi-driver/0.log" Apr 17 09:30:23.000970 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:23.000945 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-8l4wz_904ead54-94c1-4e84-9851-0c8506ed7bbc/csi-node-driver-registrar/0.log" Apr 17 09:30:23.018416 ip-10-0-137-152 kubenswrapper[2569]: I0417 09:30:23.018392 2569 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-8l4wz_904ead54-94c1-4e84-9851-0c8506ed7bbc/csi-liveness-probe/0.log"