Apr 22 14:12:21.413019 ip-10-0-136-18 systemd[1]: kubelet.service: Failed to load environment files: No such file or directory Apr 22 14:12:21.413232 ip-10-0-136-18 systemd[1]: kubelet.service: Failed to run 'start-pre' task: No such file or directory Apr 22 14:12:21.413323 ip-10-0-136-18 systemd[1]: kubelet.service: Failed with result 'resources'. Apr 22 14:12:21.413747 ip-10-0-136-18 systemd[1]: Failed to start Kubernetes Kubelet. Apr 22 14:12:31.631566 ip-10-0-136-18 systemd[1]: kubelet.service: Failed to schedule restart job: Unit crio.service not found. Apr 22 14:12:31.631588 ip-10-0-136-18 systemd[1]: kubelet.service: Failed with result 'resources'. -- Boot de3bcd84bf064ca8bbd39b10bed920bb -- Apr 22 14:14:46.435927 ip-10-0-136-18 systemd[1]: Starting Kubernetes Kubelet... Apr 22 14:14:46.783106 ip-10-0-136-18 kubenswrapper[2574]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 14:14:46.783106 ip-10-0-136-18 kubenswrapper[2574]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 22 14:14:46.783106 ip-10-0-136-18 kubenswrapper[2574]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 14:14:46.783106 ip-10-0-136-18 kubenswrapper[2574]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 22 14:14:46.783106 ip-10-0-136-18 kubenswrapper[2574]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 14:14:46.785035 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.784912 2574 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 22 14:14:46.787709 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787695 2574 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 14:14:46.787709 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787709 2574 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787714 2574 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787717 2574 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787720 2574 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787723 2574 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787727 2574 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787736 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787739 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787741 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787744 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787746 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787749 2574 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787751 2574 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787754 2574 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787756 2574 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787759 2574 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787762 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787764 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787766 2574 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787769 2574 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 14:14:46.787767 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787771 2574 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787774 2574 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787777 2574 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787779 2574 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787782 2574 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787784 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787787 2574 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787789 2574 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787791 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787794 2574 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787797 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787799 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787801 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787803 2574 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787807 2574 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787811 2574 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787814 2574 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787818 2574 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787821 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 14:14:46.788236 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787824 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787827 2574 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787830 2574 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787832 2574 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787836 2574 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787838 2574 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787841 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787843 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787846 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787848 2574 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787851 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787853 2574 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787855 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787858 2574 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787860 2574 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787864 2574 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787866 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787869 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787871 2574 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787873 2574 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 14:14:46.788689 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787876 2574 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787878 2574 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787881 2574 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787883 2574 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787885 2574 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787887 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787890 2574 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787893 2574 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787895 2574 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787898 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787901 2574 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787904 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787906 2574 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787909 2574 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787911 2574 feature_gate.go:328] unrecognized feature gate: Example Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787914 2574 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787916 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787919 2574 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787921 2574 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787923 2574 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 14:14:46.789183 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787926 2574 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787928 2574 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787930 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787933 2574 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787935 2574 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.787938 2574 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788295 2574 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788301 2574 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788303 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788307 2574 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788309 2574 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788312 2574 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788314 2574 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788317 2574 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788319 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788322 2574 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788324 2574 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788327 2574 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788329 2574 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788332 2574 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 14:14:46.789646 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788334 2574 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788336 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788339 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788341 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788343 2574 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788346 2574 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788349 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788351 2574 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788353 2574 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788356 2574 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788358 2574 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788360 2574 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788362 2574 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788365 2574 feature_gate.go:328] unrecognized feature gate: Example Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788367 2574 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788369 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788372 2574 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788374 2574 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788376 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 14:14:46.790170 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788381 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788383 2574 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788386 2574 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788388 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788391 2574 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788393 2574 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788395 2574 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788398 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788400 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788403 2574 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788405 2574 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788408 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788411 2574 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788413 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788416 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788418 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788420 2574 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788423 2574 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788425 2574 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 14:14:46.790619 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788428 2574 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788432 2574 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788435 2574 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788438 2574 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788441 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788443 2574 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788446 2574 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788449 2574 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788451 2574 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788453 2574 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788456 2574 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788458 2574 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788460 2574 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788463 2574 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788466 2574 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788469 2574 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788471 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788473 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788476 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 14:14:46.791124 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788478 2574 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788480 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788483 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788485 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788487 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788491 2574 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788493 2574 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788495 2574 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788497 2574 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788500 2574 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788502 2574 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788504 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788506 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788510 2574 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.788514 2574 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789095 2574 flags.go:64] FLAG: --address="0.0.0.0" Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789104 2574 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789111 2574 flags.go:64] FLAG: --anonymous-auth="true" Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789115 2574 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789120 2574 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 22 14:14:46.791582 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789123 2574 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789127 2574 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789131 2574 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789135 2574 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789137 2574 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789141 2574 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789144 2574 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789147 2574 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789150 2574 flags.go:64] FLAG: --cgroup-root="" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789153 2574 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789156 2574 flags.go:64] FLAG: --client-ca-file="" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789159 2574 flags.go:64] FLAG: --cloud-config="" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789161 2574 flags.go:64] FLAG: --cloud-provider="external" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789164 2574 flags.go:64] FLAG: --cluster-dns="[]" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789168 2574 flags.go:64] FLAG: --cluster-domain="" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789170 2574 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789173 2574 flags.go:64] FLAG: --config-dir="" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789176 2574 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789179 2574 flags.go:64] FLAG: --container-log-max-files="5" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789183 2574 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789186 2574 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789189 2574 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789192 2574 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789195 2574 flags.go:64] FLAG: --contention-profiling="false" Apr 22 14:14:46.792059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789198 2574 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789201 2574 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789204 2574 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789208 2574 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789212 2574 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789214 2574 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789217 2574 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789220 2574 flags.go:64] FLAG: --enable-load-reader="false" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789223 2574 flags.go:64] FLAG: --enable-server="true" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789226 2574 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789230 2574 flags.go:64] FLAG: --event-burst="100" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789233 2574 flags.go:64] FLAG: --event-qps="50" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789236 2574 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789239 2574 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789242 2574 flags.go:64] FLAG: --eviction-hard="" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789245 2574 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789248 2574 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789251 2574 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789254 2574 flags.go:64] FLAG: --eviction-soft="" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789256 2574 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789259 2574 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789262 2574 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789265 2574 flags.go:64] FLAG: --experimental-mounter-path="" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789267 2574 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789270 2574 flags.go:64] FLAG: --fail-swap-on="true" Apr 22 14:14:46.792628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789273 2574 flags.go:64] FLAG: --feature-gates="" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789276 2574 flags.go:64] FLAG: --file-check-frequency="20s" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789279 2574 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789282 2574 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789285 2574 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789287 2574 flags.go:64] FLAG: --healthz-port="10248" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789290 2574 flags.go:64] FLAG: --help="false" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789293 2574 flags.go:64] FLAG: --hostname-override="ip-10-0-136-18.ec2.internal" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789296 2574 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789298 2574 flags.go:64] FLAG: --http-check-frequency="20s" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789301 2574 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789305 2574 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789309 2574 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789312 2574 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789315 2574 flags.go:64] FLAG: --image-service-endpoint="" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789317 2574 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789320 2574 flags.go:64] FLAG: --kube-api-burst="100" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789323 2574 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789325 2574 flags.go:64] FLAG: --kube-api-qps="50" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789328 2574 flags.go:64] FLAG: --kube-reserved="" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789331 2574 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789334 2574 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789337 2574 flags.go:64] FLAG: --kubelet-cgroups="" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789339 2574 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 22 14:14:46.793223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789342 2574 flags.go:64] FLAG: --lock-file="" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789344 2574 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789347 2574 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789349 2574 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789354 2574 flags.go:64] FLAG: --log-json-split-stream="false" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789357 2574 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789359 2574 flags.go:64] FLAG: --log-text-split-stream="false" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789362 2574 flags.go:64] FLAG: --logging-format="text" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789364 2574 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789367 2574 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789370 2574 flags.go:64] FLAG: --manifest-url="" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789373 2574 flags.go:64] FLAG: --manifest-url-header="" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789377 2574 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789380 2574 flags.go:64] FLAG: --max-open-files="1000000" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789384 2574 flags.go:64] FLAG: --max-pods="110" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789387 2574 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789389 2574 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789392 2574 flags.go:64] FLAG: --memory-manager-policy="None" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789395 2574 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789397 2574 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789402 2574 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789405 2574 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789412 2574 flags.go:64] FLAG: --node-status-max-images="50" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789414 2574 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789417 2574 flags.go:64] FLAG: --oom-score-adj="-999" Apr 22 14:14:46.793809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789420 2574 flags.go:64] FLAG: --pod-cidr="" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789423 2574 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789428 2574 flags.go:64] FLAG: --pod-manifest-path="" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789430 2574 flags.go:64] FLAG: --pod-max-pids="-1" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789433 2574 flags.go:64] FLAG: --pods-per-core="0" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789436 2574 flags.go:64] FLAG: --port="10250" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789439 2574 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789442 2574 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-083fdd18f781c7dfa" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789445 2574 flags.go:64] FLAG: --qos-reserved="" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789448 2574 flags.go:64] FLAG: --read-only-port="10255" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789451 2574 flags.go:64] FLAG: --register-node="true" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789453 2574 flags.go:64] FLAG: --register-schedulable="true" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789456 2574 flags.go:64] FLAG: --register-with-taints="" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789459 2574 flags.go:64] FLAG: --registry-burst="10" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789462 2574 flags.go:64] FLAG: --registry-qps="5" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789465 2574 flags.go:64] FLAG: --reserved-cpus="" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789467 2574 flags.go:64] FLAG: --reserved-memory="" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789471 2574 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789474 2574 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789476 2574 flags.go:64] FLAG: --rotate-certificates="false" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789479 2574 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789482 2574 flags.go:64] FLAG: --runonce="false" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789485 2574 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789488 2574 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789491 2574 flags.go:64] FLAG: --seccomp-default="false" Apr 22 14:14:46.794393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789494 2574 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789497 2574 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789500 2574 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789503 2574 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789505 2574 flags.go:64] FLAG: --storage-driver-password="root" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789508 2574 flags.go:64] FLAG: --storage-driver-secure="false" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789511 2574 flags.go:64] FLAG: --storage-driver-table="stats" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789513 2574 flags.go:64] FLAG: --storage-driver-user="root" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789516 2574 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789518 2574 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789521 2574 flags.go:64] FLAG: --system-cgroups="" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789524 2574 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789529 2574 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789531 2574 flags.go:64] FLAG: --tls-cert-file="" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789534 2574 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789538 2574 flags.go:64] FLAG: --tls-min-version="" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789540 2574 flags.go:64] FLAG: --tls-private-key-file="" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789543 2574 flags.go:64] FLAG: --topology-manager-policy="none" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789546 2574 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789548 2574 flags.go:64] FLAG: --topology-manager-scope="container" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789551 2574 flags.go:64] FLAG: --v="2" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789555 2574 flags.go:64] FLAG: --version="false" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789559 2574 flags.go:64] FLAG: --vmodule="" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789564 2574 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.789567 2574 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 22 14:14:46.795008 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789655 2574 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789659 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789663 2574 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789666 2574 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789668 2574 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789684 2574 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789686 2574 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789689 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789692 2574 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789701 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789705 2574 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789709 2574 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789712 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789714 2574 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789716 2574 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789719 2574 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789722 2574 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789724 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789727 2574 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789730 2574 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 14:14:46.795609 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789733 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789735 2574 feature_gate.go:328] unrecognized feature gate: Example Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789738 2574 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789740 2574 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789742 2574 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789746 2574 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789749 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789752 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789754 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789757 2574 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789759 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789761 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789764 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789767 2574 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789769 2574 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789772 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789774 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789777 2574 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789779 2574 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789782 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 14:14:46.796395 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789784 2574 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789787 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789789 2574 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789792 2574 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789794 2574 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789797 2574 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789799 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789801 2574 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789804 2574 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789806 2574 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789812 2574 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789815 2574 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789817 2574 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789819 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789822 2574 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789824 2574 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789826 2574 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789829 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789831 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789833 2574 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 14:14:46.796946 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789836 2574 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789838 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789841 2574 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789843 2574 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789846 2574 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789848 2574 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789850 2574 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789854 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789856 2574 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789859 2574 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789861 2574 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789864 2574 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789866 2574 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789868 2574 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789870 2574 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789873 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789875 2574 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789878 2574 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789881 2574 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 14:14:46.797423 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789883 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 14:14:46.797932 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789887 2574 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 14:14:46.797932 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789891 2574 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 14:14:46.797932 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789895 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 14:14:46.797932 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789897 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 14:14:46.797932 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789900 2574 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 14:14:46.797932 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.789903 2574 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 14:14:46.797932 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.790595 2574 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 14:14:46.798942 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.798924 2574 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 22 14:14:46.798979 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.798943 2574 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 22 14:14:46.799008 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.798995 2574 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 14:14:46.799008 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799000 2574 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 14:14:46.799008 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799003 2574 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 14:14:46.799008 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799006 2574 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 14:14:46.799008 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799009 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799012 2574 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799015 2574 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799017 2574 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799020 2574 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799022 2574 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799025 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799027 2574 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799030 2574 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799033 2574 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799037 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799040 2574 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799042 2574 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799045 2574 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799047 2574 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799050 2574 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799052 2574 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799054 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799057 2574 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799059 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 14:14:46.799125 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799063 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799065 2574 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799067 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799069 2574 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799072 2574 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799074 2574 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799076 2574 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799079 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799082 2574 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799084 2574 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799086 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799089 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799091 2574 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799095 2574 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799100 2574 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799103 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799106 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799109 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799111 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 14:14:46.799588 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799114 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799117 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799119 2574 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799122 2574 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799124 2574 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799127 2574 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799130 2574 feature_gate.go:328] unrecognized feature gate: Example Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799132 2574 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799135 2574 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799137 2574 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799140 2574 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799142 2574 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799144 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799147 2574 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799149 2574 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799151 2574 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799154 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799156 2574 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799159 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799161 2574 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 14:14:46.800072 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799164 2574 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799166 2574 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799169 2574 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799171 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799173 2574 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799176 2574 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799178 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799181 2574 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799184 2574 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799186 2574 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799188 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799191 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799194 2574 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799197 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799199 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799201 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799204 2574 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799206 2574 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799209 2574 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799212 2574 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 14:14:46.800539 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799214 2574 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799217 2574 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799219 2574 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.799224 2574 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799308 2574 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799313 2574 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799316 2574 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799318 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799322 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799325 2574 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799327 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799330 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799333 2574 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799335 2574 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799338 2574 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799340 2574 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 14:14:46.801130 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799343 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799345 2574 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799348 2574 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799351 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799353 2574 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799355 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799358 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799360 2574 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799362 2574 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799365 2574 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799368 2574 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799370 2574 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799372 2574 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799375 2574 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799377 2574 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799379 2574 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799382 2574 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799384 2574 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799386 2574 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799389 2574 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 14:14:46.801508 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799391 2574 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799393 2574 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799396 2574 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799398 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799401 2574 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799404 2574 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799406 2574 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799408 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799411 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799413 2574 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799415 2574 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799418 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799420 2574 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799424 2574 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799427 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799430 2574 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799433 2574 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799435 2574 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799438 2574 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 14:14:46.802014 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799440 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799443 2574 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799445 2574 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799447 2574 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799450 2574 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799453 2574 feature_gate.go:328] unrecognized feature gate: Example Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799455 2574 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799457 2574 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799460 2574 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799462 2574 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799464 2574 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799466 2574 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799468 2574 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799471 2574 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799473 2574 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799475 2574 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799478 2574 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799481 2574 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799483 2574 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 14:14:46.802453 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799485 2574 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799488 2574 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799490 2574 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799493 2574 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799495 2574 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799498 2574 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799500 2574 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799503 2574 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799507 2574 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799509 2574 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799512 2574 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799514 2574 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799516 2574 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799518 2574 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799521 2574 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:46.799523 2574 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 14:14:46.802984 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.799528 2574 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 14:14:46.803351 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.799620 2574 server.go:962] "Client rotation is on, will bootstrap in background" Apr 22 14:14:46.803414 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.803400 2574 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 22 14:14:46.804156 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.804145 2574 server.go:1019] "Starting client certificate rotation" Apr 22 14:14:46.804257 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.804239 2574 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 14:14:46.804285 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.804276 2574 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 14:14:46.824947 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.824929 2574 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 14:14:46.827204 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.827179 2574 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 14:14:46.837607 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.837589 2574 log.go:25] "Validated CRI v1 runtime API" Apr 22 14:14:46.843014 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.843000 2574 log.go:25] "Validated CRI v1 image API" Apr 22 14:14:46.847442 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.847422 2574 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 22 14:14:46.850734 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.850709 2574 fs.go:135] Filesystem UUIDs: map[66bd6157-d487-435f-94b4-0bcbe0462118:/dev/nvme0n1p3 7B77-95E7:/dev/nvme0n1p2 99ae6b9f-7c98-441b-afdb-54287e846b3c:/dev/nvme0n1p4] Apr 22 14:14:46.850806 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.850733 2574 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 22 14:14:46.856052 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.855951 2574 manager.go:217] Machine: {Timestamp:2026-04-22 14:14:46.854972439 +0000 UTC m=+0.322850161 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3205727 MemoryCapacity:32812171264 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec21c87a197f192949c0b413145b77cd SystemUUID:ec21c87a-197f-1929-49c0-b413145b77cd BootID:de3bcd84-bf06-4ca8-bbd3-9b10bed920bb Filesystems:[{Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16406085632 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16406085632 Type:vfs Inodes:4005392 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6562435072 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:18:1c:d0:2b:bb Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:18:1c:d0:2b:bb Speed:0 Mtu:9001} {Name:ovs-system MacAddress:da:b4:d3:30:29:93 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:32812171264 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:34603008 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 22 14:14:46.856052 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.856040 2574 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 22 14:14:46.856179 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.856109 2574 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 22 14:14:46.857835 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.857789 2574 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 22 14:14:46.857982 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.857837 2574 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-136-18.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 22 14:14:46.858024 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.857992 2574 topology_manager.go:138] "Creating topology manager with none policy" Apr 22 14:14:46.858024 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.858001 2574 container_manager_linux.go:306] "Creating device plugin manager" Apr 22 14:14:46.858024 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.858013 2574 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 14:14:46.858640 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.858630 2574 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 14:14:46.860283 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.860272 2574 state_mem.go:36] "Initialized new in-memory state store" Apr 22 14:14:46.860395 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.860386 2574 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 22 14:14:46.861136 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.861119 2574 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 14:14:46.862230 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.862218 2574 kubelet.go:491] "Attempting to sync node with API server" Apr 22 14:14:46.862265 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.862237 2574 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 22 14:14:46.862265 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.862249 2574 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 22 14:14:46.862265 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.862257 2574 kubelet.go:397] "Adding apiserver pod source" Apr 22 14:14:46.862265 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.862265 2574 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 22 14:14:46.863127 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.863115 2574 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 14:14:46.863178 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.863133 2574 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 14:14:46.865342 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.865327 2574 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 22 14:14:46.866992 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.866979 2574 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 22 14:14:46.868042 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868031 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 22 14:14:46.868077 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868048 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 22 14:14:46.868077 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868054 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 22 14:14:46.868077 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868058 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 22 14:14:46.868077 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868063 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 22 14:14:46.868077 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868069 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 22 14:14:46.868077 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868074 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 22 14:14:46.868077 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868079 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 22 14:14:46.868241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868086 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 22 14:14:46.868241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868091 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 22 14:14:46.868241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868105 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 22 14:14:46.868241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868114 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 22 14:14:46.868796 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868784 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 22 14:14:46.868829 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.868800 2574 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 22 14:14:46.872228 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.872215 2574 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 22 14:14:46.872298 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.872253 2574 server.go:1295] "Started kubelet" Apr 22 14:14:46.872365 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.872340 2574 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 22 14:14:46.872819 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.872404 2574 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 22 14:14:46.872874 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.872853 2574 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 22 14:14:46.873038 ip-10-0-136-18 systemd[1]: Started Kubernetes Kubelet. Apr 22 14:14:46.873920 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.873870 2574 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 22 14:14:46.874731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.874715 2574 server.go:317] "Adding debug handlers to kubelet server" Apr 22 14:14:46.876249 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.876230 2574 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-136-18.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 22 14:14:46.876503 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.876478 2574 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-136-18.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 22 14:14:46.876554 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.876478 2574 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 22 14:14:46.879331 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.879314 2574 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 22 14:14:46.879703 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.879687 2574 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 22 14:14:46.880539 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.880476 2574 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 22 14:14:46.880539 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.880494 2574 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 22 14:14:46.880703 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.880559 2574 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 22 14:14:46.880703 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.880594 2574 reconstruct.go:97] "Volume reconstruction finished" Apr 22 14:14:46.880703 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.880598 2574 reconciler.go:26] "Reconciler: start to sync state" Apr 22 14:14:46.881603 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.881516 2574 factory.go:55] Registering systemd factory Apr 22 14:14:46.881603 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.881541 2574 factory.go:223] Registration of the systemd container factory successfully Apr 22 14:14:46.881770 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.881717 2574 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 22 14:14:46.881876 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.881854 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:46.882817 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.882794 2574 factory.go:153] Registering CRI-O factory Apr 22 14:14:46.882887 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.882821 2574 factory.go:223] Registration of the crio container factory successfully Apr 22 14:14:46.882943 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.882890 2574 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 22 14:14:46.882943 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.882911 2574 factory.go:103] Registering Raw factory Apr 22 14:14:46.882943 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.882937 2574 manager.go:1196] Started watching for new ooms in manager Apr 22 14:14:46.883268 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.883235 2574 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 22 14:14:46.884132 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.884108 2574 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"ip-10-0-136-18.ec2.internal\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Apr 22 14:14:46.885153 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.885135 2574 manager.go:319] Starting recovery of all containers Apr 22 14:14:46.886828 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.883033 2574 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-136-18.ec2.internal.18a8b3621707fe18 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-136-18.ec2.internal,UID:ip-10-0-136-18.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-10-0-136-18.ec2.internal,},FirstTimestamp:2026-04-22 14:14:46.872227352 +0000 UTC m=+0.340105072,LastTimestamp:2026-04-22 14:14:46.872227352 +0000 UTC m=+0.340105072,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-136-18.ec2.internal,}" Apr 22 14:14:46.895788 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.895695 2574 manager.go:324] Recovery completed Apr 22 14:14:46.899829 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.899817 2574 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 14:14:46.902092 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.902075 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientMemory" Apr 22 14:14:46.902158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.902101 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 14:14:46.902158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.902110 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientPID" Apr 22 14:14:46.902512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.902500 2574 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 22 14:14:46.902512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.902510 2574 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 22 14:14:46.902591 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.902524 2574 state_mem.go:36] "Initialized new in-memory state store" Apr 22 14:14:46.903687 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.903617 2574 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-136-18.ec2.internal.18a8b36218cfa98e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-136-18.ec2.internal,UID:ip-10-0-136-18.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node ip-10-0-136-18.ec2.internal status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:ip-10-0-136-18.ec2.internal,},FirstTimestamp:2026-04-22 14:14:46.902090126 +0000 UTC m=+0.369967847,LastTimestamp:2026-04-22 14:14:46.902090126 +0000 UTC m=+0.369967847,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-136-18.ec2.internal,}" Apr 22 14:14:46.905666 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.905653 2574 policy_none.go:49] "None policy: Start" Apr 22 14:14:46.905742 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.905702 2574 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 22 14:14:46.905742 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.905714 2574 state_mem.go:35] "Initializing new in-memory state store" Apr 22 14:14:46.907897 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.907882 2574 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-lnj2j" Apr 22 14:14:46.912968 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.912908 2574 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-136-18.ec2.internal.18a8b36218cfe4fa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-136-18.ec2.internal,UID:ip-10-0-136-18.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node ip-10-0-136-18.ec2.internal status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:ip-10-0-136-18.ec2.internal,},FirstTimestamp:2026-04-22 14:14:46.902105338 +0000 UTC m=+0.369983059,LastTimestamp:2026-04-22 14:14:46.902105338 +0000 UTC m=+0.369983059,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-136-18.ec2.internal,}" Apr 22 14:14:46.915901 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.915876 2574 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-lnj2j" Apr 22 14:14:46.939210 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.939196 2574 manager.go:341] "Starting Device Plugin manager" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.939228 2574 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.939241 2574 server.go:85] "Starting device plugin registration server" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.939473 2574 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.939484 2574 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.939561 2574 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.939728 2574 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.939736 2574 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.940185 2574 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 22 14:14:46.961093 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.940210 2574 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:46.981471 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.981451 2574 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 22 14:14:46.982759 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.982744 2574 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 22 14:14:46.982823 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.982776 2574 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 22 14:14:46.982823 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.982791 2574 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 22 14:14:46.982823 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.982798 2574 kubelet.go:2451] "Starting kubelet main sync loop" Apr 22 14:14:46.982977 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:46.982827 2574 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 22 14:14:46.985215 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:46.985198 2574 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 14:14:47.040485 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.040436 2574 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 14:14:47.042814 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.042798 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientMemory" Apr 22 14:14:47.042909 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.042831 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 14:14:47.042909 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.042849 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientPID" Apr 22 14:14:47.042909 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.042879 2574 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.054545 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.054528 2574 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.054601 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.054547 2574 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-136-18.ec2.internal\": node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.064084 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.064066 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.083826 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.083806 2574 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal"] Apr 22 14:14:47.083902 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.083876 2574 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 14:14:47.085747 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.085735 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientMemory" Apr 22 14:14:47.085807 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.085764 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 14:14:47.085807 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.085795 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientPID" Apr 22 14:14:47.088097 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088085 2574 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 14:14:47.088245 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088231 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.088303 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088265 2574 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 14:14:47.088787 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088769 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientMemory" Apr 22 14:14:47.088881 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088797 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 14:14:47.088881 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088809 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientPID" Apr 22 14:14:47.088881 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088798 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientMemory" Apr 22 14:14:47.088881 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088878 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 14:14:47.089011 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.088888 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientPID" Apr 22 14:14:47.091126 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.091113 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.091183 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.091135 2574 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 14:14:47.091711 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.091697 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientMemory" Apr 22 14:14:47.091788 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.091719 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 14:14:47.091788 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.091728 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeHasSufficientPID" Apr 22 14:14:47.121246 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.121228 2574 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-136-18.ec2.internal\" not found" node="ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.125300 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.125286 2574 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-136-18.ec2.internal\" not found" node="ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.164886 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.164868 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.181973 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.181953 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/e45c97e69339b1e77c7fd7ba3167751b-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal\" (UID: \"e45c97e69339b1e77c7fd7ba3167751b\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.182051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.181979 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e45c97e69339b1e77c7fd7ba3167751b-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal\" (UID: \"e45c97e69339b1e77c7fd7ba3167751b\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.182051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.181994 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/17be8cc2ca560f8da5581c8d587a8582-config\") pod \"kube-apiserver-proxy-ip-10-0-136-18.ec2.internal\" (UID: \"17be8cc2ca560f8da5581c8d587a8582\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.265136 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.265118 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.282393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.282371 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/e45c97e69339b1e77c7fd7ba3167751b-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal\" (UID: \"e45c97e69339b1e77c7fd7ba3167751b\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.282488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.282402 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e45c97e69339b1e77c7fd7ba3167751b-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal\" (UID: \"e45c97e69339b1e77c7fd7ba3167751b\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.282488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.282417 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/17be8cc2ca560f8da5581c8d587a8582-config\") pod \"kube-apiserver-proxy-ip-10-0-136-18.ec2.internal\" (UID: \"17be8cc2ca560f8da5581c8d587a8582\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.282488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.282463 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/17be8cc2ca560f8da5581c8d587a8582-config\") pod \"kube-apiserver-proxy-ip-10-0-136-18.ec2.internal\" (UID: \"17be8cc2ca560f8da5581c8d587a8582\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.282488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.282471 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/e45c97e69339b1e77c7fd7ba3167751b-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal\" (UID: \"e45c97e69339b1e77c7fd7ba3167751b\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.282488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.282475 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e45c97e69339b1e77c7fd7ba3167751b-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal\" (UID: \"e45c97e69339b1e77c7fd7ba3167751b\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.365744 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.365697 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.423182 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.423158 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.427686 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.427652 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" Apr 22 14:14:47.466287 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.466258 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.566768 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.566747 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.667227 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.667208 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.757653 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.757629 2574 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 14:14:47.768051 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.768030 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.804469 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.804446 2574 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 22 14:14:47.804895 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.804588 2574 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 22 14:14:47.804895 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.804614 2574 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 22 14:14:47.869031 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.869005 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.879974 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.879956 2574 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 22 14:14:47.895215 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.895192 2574 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 14:14:47.917766 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.917710 2574 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-21 14:09:46 +0000 UTC" deadline="2027-10-28 07:07:24.277864798 +0000 UTC" Apr 22 14:14:47.917766 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.917738 2574 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="13288h52m36.360129298s" Apr 22 14:14:47.939616 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.939595 2574 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-lvlf6" Apr 22 14:14:47.950809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.950794 2574 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-lvlf6" Apr 22 14:14:47.953691 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:47.953628 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17be8cc2ca560f8da5581c8d587a8582.slice/crio-dac84ba4f1223c889ce52a60e3e9601d94fe32d48f5e4417c973490bc30fe7b5 WatchSource:0}: Error finding container dac84ba4f1223c889ce52a60e3e9601d94fe32d48f5e4417c973490bc30fe7b5: Status 404 returned error can't find the container with id dac84ba4f1223c889ce52a60e3e9601d94fe32d48f5e4417c973490bc30fe7b5 Apr 22 14:14:47.954018 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:47.953998 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode45c97e69339b1e77c7fd7ba3167751b.slice/crio-c8b91c153d6c2a9c03b1b03e03119b475683bc777acee36f6aa51b2d76096e0a WatchSource:0}: Error finding container c8b91c153d6c2a9c03b1b03e03119b475683bc777acee36f6aa51b2d76096e0a: Status 404 returned error can't find the container with id c8b91c153d6c2a9c03b1b03e03119b475683bc777acee36f6aa51b2d76096e0a Apr 22 14:14:47.957512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.957500 2574 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 14:14:47.969152 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:47.969129 2574 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-136-18.ec2.internal\" not found" Apr 22 14:14:47.985114 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.985075 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" event={"ID":"e45c97e69339b1e77c7fd7ba3167751b","Type":"ContainerStarted","Data":"c8b91c153d6c2a9c03b1b03e03119b475683bc777acee36f6aa51b2d76096e0a"} Apr 22 14:14:47.986024 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:47.985998 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" event={"ID":"17be8cc2ca560f8da5581c8d587a8582","Type":"ContainerStarted","Data":"dac84ba4f1223c889ce52a60e3e9601d94fe32d48f5e4417c973490bc30fe7b5"} Apr 22 14:14:48.064361 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.064340 2574 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 14:14:48.080826 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.080798 2574 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" Apr 22 14:14:48.093682 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.093663 2574 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 14:14:48.094609 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.094598 2574 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" Apr 22 14:14:48.104000 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.103982 2574 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 14:14:48.365255 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.365193 2574 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 14:14:48.863054 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.863030 2574 apiserver.go:52] "Watching apiserver" Apr 22 14:14:48.871125 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.871104 2574 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 22 14:14:48.872736 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.872715 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-68xhd","kube-system/konnectivity-agent-f42dv","openshift-cluster-node-tuning-operator/tuned-kg26l","openshift-multus/multus-additional-cni-plugins-25xzb","openshift-network-operator/iptables-alerter-9dhqr","openshift-ovn-kubernetes/ovnkube-node-gpjz8","kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7","openshift-dns/node-resolver-d5pcf","openshift-image-registry/node-ca-dx5kd","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal","openshift-multus/multus-hwrgf","openshift-multus/network-metrics-daemon-p65qg"] Apr 22 14:14:48.877249 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.877228 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.879305 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.879285 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:48.879402 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.879342 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-ppfc2\"" Apr 22 14:14:48.879402 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.879384 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 22 14:14:48.879650 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.879631 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 22 14:14:48.879766 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.879752 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 22 14:14:48.881269 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.881253 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 22 14:14:48.881923 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.881905 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-4n44t\"" Apr 22 14:14:48.881923 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.881905 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 22 14:14:48.883450 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.883079 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.885158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.885138 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 22 14:14:48.885282 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.885261 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-2b9zp\"" Apr 22 14:14:48.885350 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.885306 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 22 14:14:48.886171 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.886042 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.886315 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.886299 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:48.888472 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.888454 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 22 14:14:48.888551 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.888494 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-dzcwp\"" Apr 22 14:14:48.888596 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.888571 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 22 14:14:48.888972 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.888952 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-system-cni-dir\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.889066 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.888987 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-os-release\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.889066 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889023 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.889066 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889033 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-systemd\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889066 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889059 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-lib-modules\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889081 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-kubernetes\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889106 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889117 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/67d6b18b-35ca-4116-935f-5107cc660741-etc-tuned\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889139 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/67d6b18b-35ca-4116-935f-5107cc660741-tmp\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889162 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6hbt\" (UniqueName: \"kubernetes.io/projected/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-kube-api-access-z6hbt\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889191 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3688a996-ab5a-4d60-b8e0-19b8a75bae48-host\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889214 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-modprobe-d\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889246 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889310 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-97skw\"" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889247 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysconfig\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889427 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysctl-conf\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889449 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-sys\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889475 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3688a996-ab5a-4d60-b8e0-19b8a75bae48-serviceca\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889520 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-host\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889543 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cnibin\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889569 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889581 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cni-binary-copy\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.889626 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889614 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889638 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889666 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889716 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/4cf8ecc7-e6d0-4d13-9d56-8e329378e84a-konnectivity-ca\") pod \"konnectivity-agent-f42dv\" (UID: \"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a\") " pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889742 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-var-lib-kubelet\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889764 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889793 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysctl-d\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889806 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889815 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-run\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889845 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nljmh\" (UniqueName: \"kubernetes.io/projected/3688a996-ab5a-4d60-b8e0-19b8a75bae48-kube-api-access-nljmh\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889881 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889887 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/4cf8ecc7-e6d0-4d13-9d56-8e329378e84a-agent-certs\") pod \"konnectivity-agent-f42dv\" (UID: \"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a\") " pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:48.890100 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.889909 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwm7p\" (UniqueName: \"kubernetes.io/projected/67d6b18b-35ca-4116-935f-5107cc660741-kube-api-access-rwm7p\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.890704 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.890686 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 22 14:14:48.891009 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.890991 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-sncpp\"" Apr 22 14:14:48.891306 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.891289 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.891631 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.891614 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 22 14:14:48.891733 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.891695 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 22 14:14:48.891902 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.891885 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 22 14:14:48.892067 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.892051 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 22 14:14:48.892155 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.892072 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 22 14:14:48.893075 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.893057 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 22 14:14:48.893941 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.893912 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 22 14:14:48.893941 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.893926 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-x9sz9\"" Apr 22 14:14:48.893941 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.893936 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 22 14:14:48.895890 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.895872 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:48.897619 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.897584 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 22 14:14:48.897732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.897589 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 22 14:14:48.897855 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.897809 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-n4pfj\"" Apr 22 14:14:48.898666 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.898194 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:48.898666 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.898279 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.898666 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:48.898303 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:14:48.900049 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.900028 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 22 14:14:48.900049 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.900046 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-hqgwb\"" Apr 22 14:14:48.900543 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.900526 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:48.900613 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:48.900587 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:14:48.951382 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.951347 2574 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 14:09:47 +0000 UTC" deadline="2028-01-22 16:24:24.882376272 +0000 UTC" Apr 22 14:14:48.951382 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.951381 2574 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="15362h9m35.93099724s" Apr 22 14:14:48.981351 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.981327 2574 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 22 14:14:48.990441 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990414 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.990522 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990452 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cd5f393a-8e7d-4fec-ba9c-362d5579a518-hosts-file\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:48.990522 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990481 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/67d6b18b-35ca-4116-935f-5107cc660741-etc-tuned\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.990608 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990522 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/67d6b18b-35ca-4116-935f-5107cc660741-tmp\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.990608 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990557 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-k8s-cni-cncf-io\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.990608 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990583 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-var-lib-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.990735 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990606 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-etc-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.990735 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990634 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3688a996-ab5a-4d60-b8e0-19b8a75bae48-host\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.990735 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990653 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysconfig\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.990735 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990690 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysctl-conf\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.990735 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990713 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-hostroot\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990736 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-ovn\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990757 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-cni-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990780 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-run-netns\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990805 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovn-node-metrics-cert\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990809 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3688a996-ab5a-4d60-b8e0-19b8a75bae48-host\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990808 2574 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990861 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysconfig\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.990914 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990889 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-socket-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990913 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/6a01b9d7-2c6c-4443-a988-d06252bd1a67-iptables-alerter-script\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990941 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3688a996-ab5a-4d60-b8e0-19b8a75bae48-serviceca\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990954 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysctl-conf\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990967 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/cd5f393a-8e7d-4fec-ba9c-362d5579a518-tmp-dir\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.990990 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cni-binary-copy\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991033 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991057 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/4cf8ecc7-e6d0-4d13-9d56-8e329378e84a-konnectivity-ca\") pod \"konnectivity-agent-f42dv\" (UID: \"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a\") " pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991079 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-var-lib-kubelet\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991104 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-system-cni-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991128 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a01b9d7-2c6c-4443-a988-d06252bd1a67-host-slash\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991151 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ec55872-0394-47c2-9108-d21ab221b5f8-cni-binary-copy\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991172 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-slash\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.991201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991193 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-node-log\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991217 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-etc-selinux\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991241 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-cni-bin\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991263 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-etc-kubernetes\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991284 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-systemd\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991326 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovnkube-config\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991348 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-registration-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991364 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3688a996-ab5a-4d60-b8e0-19b8a75bae48-serviceca\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991371 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-sys-fs\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991394 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9hd2\" (UniqueName: \"kubernetes.io/projected/6a01b9d7-2c6c-4443-a988-d06252bd1a67-kube-api-access-d9hd2\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991425 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-os-release\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991487 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cni-binary-copy\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991495 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-var-lib-kubelet\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991490 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-lib-modules\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991532 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-netns\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991556 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-conf-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991597 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-kubernetes\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.991652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991624 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-log-socket\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991651 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwmlb\" (UniqueName: \"kubernetes.io/projected/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-kube-api-access-mwmlb\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991746 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pzdk\" (UniqueName: \"kubernetes.io/projected/cd5f393a-8e7d-4fec-ba9c-362d5579a518-kube-api-access-5pzdk\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991909 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-z6hbt\" (UniqueName: \"kubernetes.io/projected/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-kube-api-access-z6hbt\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991918 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991932 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-modprobe-d\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991952 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-sys\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991973 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-os-release\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991991 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-os-release\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.991993 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-cni-multus\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992002 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/4cf8ecc7-e6d0-4d13-9d56-8e329378e84a-konnectivity-ca\") pod \"konnectivity-agent-f42dv\" (UID: \"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a\") " pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992028 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-systemd-units\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992071 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-sys\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992080 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-kubernetes\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992098 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-lib-modules\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992106 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-cni-bin\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992131 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-env-overrides\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.992248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992155 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-modprobe-d\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992158 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-host\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992190 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-host\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992216 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-multus-certs\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992233 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992262 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-device-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992293 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cnibin\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992330 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cnibin\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992365 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992430 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrsm7\" (UniqueName: \"kubernetes.io/projected/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-kube-api-access-nrsm7\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992503 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-run-ovn-kubernetes\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992521 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992529 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-cni-netd\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992553 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovnkube-script-lib\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992579 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldss\" (UniqueName: \"kubernetes.io/projected/3cff0f27-e53b-4843-a2ed-7e885a500c4a-kube-api-access-bldss\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992609 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-kubelet-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:48.993007 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992636 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992660 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysctl-d\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992698 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-run\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992722 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-cnibin\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992744 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-daemon-config\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992767 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992793 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nljmh\" (UniqueName: \"kubernetes.io/projected/3688a996-ab5a-4d60-b8e0-19b8a75bae48-kube-api-access-nljmh\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992821 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/4cf8ecc7-e6d0-4d13-9d56-8e329378e84a-agent-certs\") pod \"konnectivity-agent-f42dv\" (UID: \"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a\") " pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992845 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rwm7p\" (UniqueName: \"kubernetes.io/projected/67d6b18b-35ca-4116-935f-5107cc660741-kube-api-access-rwm7p\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992872 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992895 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-kubelet\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992921 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-system-cni-dir\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992944 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-systemd\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992968 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-socket-dir-parent\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.992994 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s79mr\" (UniqueName: \"kubernetes.io/projected/4ec55872-0394-47c2-9108-d21ab221b5f8-kube-api-access-s79mr\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.993017 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-kubelet\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.993027 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.993731 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.993103 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-system-cni-dir\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:48.994395 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.993156 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-systemd\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.994395 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.993309 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-etc-sysctl-d\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.994395 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.993351 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d6b18b-35ca-4116-935f-5107cc660741-run\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.994495 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.994462 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/67d6b18b-35ca-4116-935f-5107cc660741-etc-tuned\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.994643 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.994629 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/67d6b18b-35ca-4116-935f-5107cc660741-tmp\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:48.995625 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:48.995609 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/4cf8ecc7-e6d0-4d13-9d56-8e329378e84a-agent-certs\") pod \"konnectivity-agent-f42dv\" (UID: \"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a\") " pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:49.005375 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.005356 2574 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 14:14:49.007431 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.007407 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nljmh\" (UniqueName: \"kubernetes.io/projected/3688a996-ab5a-4d60-b8e0-19b8a75bae48-kube-api-access-nljmh\") pod \"node-ca-dx5kd\" (UID: \"3688a996-ab5a-4d60-b8e0-19b8a75bae48\") " pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:49.007431 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.007424 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6hbt\" (UniqueName: \"kubernetes.io/projected/cc6913ef-cc9f-4a41-971f-4e5dda9212fd-kube-api-access-z6hbt\") pod \"multus-additional-cni-plugins-25xzb\" (UID: \"cc6913ef-cc9f-4a41-971f-4e5dda9212fd\") " pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:49.007577 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.007550 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwm7p\" (UniqueName: \"kubernetes.io/projected/67d6b18b-35ca-4116-935f-5107cc660741-kube-api-access-rwm7p\") pod \"tuned-kg26l\" (UID: \"67d6b18b-35ca-4116-935f-5107cc660741\") " pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:49.093602 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093574 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-cni-bin\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.093776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093611 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-etc-kubernetes\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.093776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093641 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-systemd\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.093776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093701 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovnkube-config\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.093776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093726 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-registration-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.093776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093725 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-cni-bin\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.093776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093747 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-systemd\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.093776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093760 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-sys-fs\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093835 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-registration-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093852 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d9hd2\" (UniqueName: \"kubernetes.io/projected/6a01b9d7-2c6c-4443-a988-d06252bd1a67-kube-api-access-d9hd2\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093764 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-etc-kubernetes\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093819 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-sys-fs\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093884 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-netns\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093902 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-conf-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093917 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-log-socket\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093937 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mwmlb\" (UniqueName: \"kubernetes.io/projected/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-kube-api-access-mwmlb\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093959 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5pzdk\" (UniqueName: \"kubernetes.io/projected/cd5f393a-8e7d-4fec-ba9c-362d5579a518-kube-api-access-5pzdk\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093990 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-log-socket\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.093989 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-conf-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094059 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094048 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-netns\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094106 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-os-release\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094142 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-cni-multus\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094160 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-systemd-units\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094187 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-cni-bin\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094197 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-cni-multus\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094209 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-env-overrides\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094234 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-multus-certs\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094237 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-systemd-units\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094233 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-os-release\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094257 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094278 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-device-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094279 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-cni-bin\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094284 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovnkube-config\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094279 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-multus-certs\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094323 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-device-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094343 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094373 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nrsm7\" (UniqueName: \"kubernetes.io/projected/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-kube-api-access-nrsm7\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:49.094575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094401 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-run-ovn-kubernetes\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094423 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-cni-netd\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094441 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovnkube-script-lib\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094474 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bldss\" (UniqueName: \"kubernetes.io/projected/3cff0f27-e53b-4843-a2ed-7e885a500c4a-kube-api-access-bldss\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094501 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-kubelet-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094518 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-cnibin\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094533 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-daemon-config\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094548 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.094624 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094633 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-env-overrides\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094752 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-run-ovn-kubernetes\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.094749 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:14:49.594692206 +0000 UTC m=+3.062569928 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094810 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-cni-netd\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094814 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-cnibin\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094871 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-kubelet-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094896 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094914 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-kubelet\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.095418 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094931 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-socket-dir-parent\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094945 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-s79mr\" (UniqueName: \"kubernetes.io/projected/4ec55872-0394-47c2-9108-d21ab221b5f8-kube-api-access-s79mr\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094961 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-kubelet\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094977 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.094993 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cd5f393a-8e7d-4fec-ba9c-362d5579a518-hosts-file\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095012 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-k8s-cni-cncf-io\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095026 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-var-lib-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095040 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-etc-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095057 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-hostroot\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095070 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-ovn\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095086 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-cni-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095099 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-run-netns\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095128 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovn-node-metrics-cert\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095143 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-socket-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095157 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/6a01b9d7-2c6c-4443-a988-d06252bd1a67-iptables-alerter-script\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095180 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/cd5f393a-8e7d-4fec-ba9c-362d5579a518-tmp-dir\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095198 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-system-cni-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095203 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovnkube-script-lib\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095213 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a01b9d7-2c6c-4443-a988-d06252bd1a67-host-slash\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095234 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a01b9d7-2c6c-4443-a988-d06252bd1a67-host-slash\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095237 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ec55872-0394-47c2-9108-d21ab221b5f8-cni-binary-copy\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095256 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-slash\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095270 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-socket-dir-parent\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095271 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-node-log\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095292 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-etc-selinux\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095296 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-node-log\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095318 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-daemon-config\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095381 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-slash\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095450 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-multus-cni-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095481 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-run-netns\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095496 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-var-lib-kubelet\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095536 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-kubelet\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095575 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095621 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-run-ovn\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095625 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-host-run-k8s-cni-cncf-io\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.096732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095655 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-etc-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095664 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cff0f27-e53b-4843-a2ed-7e885a500c4a-var-lib-openvswitch\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095709 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-hostroot\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095718 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/cd5f393a-8e7d-4fec-ba9c-362d5579a518-hosts-file\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095804 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-socket-dir\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095824 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ec55872-0394-47c2-9108-d21ab221b5f8-cni-binary-copy\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095871 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec55872-0394-47c2-9108-d21ab221b5f8-system-cni-dir\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.095939 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-etc-selinux\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.096002 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/cd5f393a-8e7d-4fec-ba9c-362d5579a518-tmp-dir\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:49.097302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.096258 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/6a01b9d7-2c6c-4443-a988-d06252bd1a67-iptables-alerter-script\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:49.097900 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.097874 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cff0f27-e53b-4843-a2ed-7e885a500c4a-ovn-node-metrics-cert\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.107183 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.107152 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 14:14:49.107183 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.107178 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 14:14:49.107350 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.107192 2574 projected.go:194] Error preparing data for projected volume kube-api-access-85plq for pod openshift-network-diagnostics/network-check-target-68xhd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:49.107350 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.107232 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-s79mr\" (UniqueName: \"kubernetes.io/projected/4ec55872-0394-47c2-9108-d21ab221b5f8-kube-api-access-s79mr\") pod \"multus-hwrgf\" (UID: \"4ec55872-0394-47c2-9108-d21ab221b5f8\") " pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.107350 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.107299 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq podName:f67ee112-d296-44c7-9c63-b48da0c3bc44 nodeName:}" failed. No retries permitted until 2026-04-22 14:14:49.607283258 +0000 UTC m=+3.075160987 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-85plq" (UniqueName: "kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq") pod "network-check-target-68xhd" (UID: "f67ee112-d296-44c7-9c63-b48da0c3bc44") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:49.109620 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.109564 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwmlb\" (UniqueName: \"kubernetes.io/projected/9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7-kube-api-access-mwmlb\") pod \"aws-ebs-csi-driver-node-nnps7\" (UID: \"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.109620 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.109584 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pzdk\" (UniqueName: \"kubernetes.io/projected/cd5f393a-8e7d-4fec-ba9c-362d5579a518-kube-api-access-5pzdk\") pod \"node-resolver-d5pcf\" (UID: \"cd5f393a-8e7d-4fec-ba9c-362d5579a518\") " pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:49.109851 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.109829 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrsm7\" (UniqueName: \"kubernetes.io/projected/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-kube-api-access-nrsm7\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:49.109955 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.109932 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9hd2\" (UniqueName: \"kubernetes.io/projected/6a01b9d7-2c6c-4443-a988-d06252bd1a67-kube-api-access-d9hd2\") pod \"iptables-alerter-9dhqr\" (UID: \"6a01b9d7-2c6c-4443-a988-d06252bd1a67\") " pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:49.110038 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.110010 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldss\" (UniqueName: \"kubernetes.io/projected/3cff0f27-e53b-4843-a2ed-7e885a500c4a-kube-api-access-bldss\") pod \"ovnkube-node-gpjz8\" (UID: \"3cff0f27-e53b-4843-a2ed-7e885a500c4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.190660 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.190635 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dx5kd" Apr 22 14:14:49.197448 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.197427 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:14:49.205041 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.205019 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-kg26l" Apr 22 14:14:49.210868 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.210848 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-9dhqr" Apr 22 14:14:49.217385 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.217366 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-25xzb" Apr 22 14:14:49.224015 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.223993 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:14:49.229735 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.229715 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" Apr 22 14:14:49.236257 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.236241 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-d5pcf" Apr 22 14:14:49.240773 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.240757 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hwrgf" Apr 22 14:14:49.599443 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.599415 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:49.599607 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.599569 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:49.599648 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.599631 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:14:50.59961415 +0000 UTC m=+4.067491872 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:49.700330 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.700306 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:49.700431 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.700412 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 14:14:49.700431 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.700424 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 14:14:49.700431 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.700431 2574 projected.go:194] Error preparing data for projected volume kube-api-access-85plq for pod openshift-network-diagnostics/network-check-target-68xhd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:49.700522 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.700473 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq podName:f67ee112-d296-44c7-9c63-b48da0c3bc44 nodeName:}" failed. No retries permitted until 2026-04-22 14:14:50.700461957 +0000 UTC m=+4.168339678 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-85plq" (UniqueName: "kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq") pod "network-check-target-68xhd" (UID: "f67ee112-d296-44c7-9c63-b48da0c3bc44") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:49.710960 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.710933 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a01b9d7_2c6c_4443_a988_d06252bd1a67.slice/crio-e2dd320066fd2e6e65b4e4147d0bcb2402eb998363d359ab1da23b22edf02be1 WatchSource:0}: Error finding container e2dd320066fd2e6e65b4e4147d0bcb2402eb998363d359ab1da23b22edf02be1: Status 404 returned error can't find the container with id e2dd320066fd2e6e65b4e4147d0bcb2402eb998363d359ab1da23b22edf02be1 Apr 22 14:14:49.712329 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.712268 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd5f393a_8e7d_4fec_ba9c_362d5579a518.slice/crio-ab56534be10c77c8bfddbec0f316bacfd80299fad5163550786fd0b2ade6a8ad WatchSource:0}: Error finding container ab56534be10c77c8bfddbec0f316bacfd80299fad5163550786fd0b2ade6a8ad: Status 404 returned error can't find the container with id ab56534be10c77c8bfddbec0f316bacfd80299fad5163550786fd0b2ade6a8ad Apr 22 14:14:49.714497 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.714103 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ec55872_0394_47c2_9108_d21ab221b5f8.slice/crio-e7258a9536fa7c4b2113cf782b5e3255f6f2e91bbd32169a8ab8d02dc688f022 WatchSource:0}: Error finding container e7258a9536fa7c4b2113cf782b5e3255f6f2e91bbd32169a8ab8d02dc688f022: Status 404 returned error can't find the container with id e7258a9536fa7c4b2113cf782b5e3255f6f2e91bbd32169a8ab8d02dc688f022 Apr 22 14:14:49.715456 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.715400 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc6913ef_cc9f_4a41_971f_4e5dda9212fd.slice/crio-a55e5b6ed4f633aae8a7620b48637270cb03a810e8bd2bd7ea42087b8041441a WatchSource:0}: Error finding container a55e5b6ed4f633aae8a7620b48637270cb03a810e8bd2bd7ea42087b8041441a: Status 404 returned error can't find the container with id a55e5b6ed4f633aae8a7620b48637270cb03a810e8bd2bd7ea42087b8041441a Apr 22 14:14:49.718898 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.718875 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cf8ecc7_e6d0_4d13_9d56_8e329378e84a.slice/crio-71e7b6a66d7daf60efa3c067e4b33fcf0aac15affa1e8130d90a376a6f50476a WatchSource:0}: Error finding container 71e7b6a66d7daf60efa3c067e4b33fcf0aac15affa1e8130d90a376a6f50476a: Status 404 returned error can't find the container with id 71e7b6a66d7daf60efa3c067e4b33fcf0aac15affa1e8130d90a376a6f50476a Apr 22 14:14:49.721021 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.720996 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3688a996_ab5a_4d60_b8e0_19b8a75bae48.slice/crio-382bb28e643eacb242929de31952957dcf2eb3c4c3ed64fb99952ff4bd88f367 WatchSource:0}: Error finding container 382bb28e643eacb242929de31952957dcf2eb3c4c3ed64fb99952ff4bd88f367: Status 404 returned error can't find the container with id 382bb28e643eacb242929de31952957dcf2eb3c4c3ed64fb99952ff4bd88f367 Apr 22 14:14:49.721277 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.721258 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67d6b18b_35ca_4116_935f_5107cc660741.slice/crio-430190a23842cb6b4d20224c7fd9a02a2714a026da84e10f6e36daf0a7ba94ea WatchSource:0}: Error finding container 430190a23842cb6b4d20224c7fd9a02a2714a026da84e10f6e36daf0a7ba94ea: Status 404 returned error can't find the container with id 430190a23842cb6b4d20224c7fd9a02a2714a026da84e10f6e36daf0a7ba94ea Apr 22 14:14:49.723996 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.723927 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cff0f27_e53b_4843_a2ed_7e885a500c4a.slice/crio-1919e838e2af507f4d551cf3abf732dc1c8609f4777b65fbea515fc4cd484d1a WatchSource:0}: Error finding container 1919e838e2af507f4d551cf3abf732dc1c8609f4777b65fbea515fc4cd484d1a: Status 404 returned error can't find the container with id 1919e838e2af507f4d551cf3abf732dc1c8609f4777b65fbea515fc4cd484d1a Apr 22 14:14:49.725259 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:14:49.724375 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ca5ec93_ab8a_4af8_a32b_6ef0c190fcf7.slice/crio-41e0d28bc0a834f665267a408250a992e586c0744d7b15ac1cde80ec60e05a21 WatchSource:0}: Error finding container 41e0d28bc0a834f665267a408250a992e586c0744d7b15ac1cde80ec60e05a21: Status 404 returned error can't find the container with id 41e0d28bc0a834f665267a408250a992e586c0744d7b15ac1cde80ec60e05a21 Apr 22 14:14:49.888342 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.888218 2574 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 14:14:49.951922 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.951890 2574 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 14:09:47 +0000 UTC" deadline="2027-12-25 10:27:49.802535589 +0000 UTC" Apr 22 14:14:49.951922 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.951922 2574 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14684h12m59.850616033s" Apr 22 14:14:49.983149 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.983128 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:49.983256 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:49.983239 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:14:49.989483 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.989428 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hwrgf" event={"ID":"4ec55872-0394-47c2-9108-d21ab221b5f8","Type":"ContainerStarted","Data":"e7258a9536fa7c4b2113cf782b5e3255f6f2e91bbd32169a8ab8d02dc688f022"} Apr 22 14:14:49.991042 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.991018 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-9dhqr" event={"ID":"6a01b9d7-2c6c-4443-a988-d06252bd1a67","Type":"ContainerStarted","Data":"e2dd320066fd2e6e65b4e4147d0bcb2402eb998363d359ab1da23b22edf02be1"} Apr 22 14:14:49.992718 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.992698 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" event={"ID":"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7","Type":"ContainerStarted","Data":"41e0d28bc0a834f665267a408250a992e586c0744d7b15ac1cde80ec60e05a21"} Apr 22 14:14:49.993721 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.993702 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"1919e838e2af507f4d551cf3abf732dc1c8609f4777b65fbea515fc4cd484d1a"} Apr 22 14:14:49.994690 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.994650 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dx5kd" event={"ID":"3688a996-ab5a-4d60-b8e0-19b8a75bae48","Type":"ContainerStarted","Data":"382bb28e643eacb242929de31952957dcf2eb3c4c3ed64fb99952ff4bd88f367"} Apr 22 14:14:49.995530 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.995514 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-d5pcf" event={"ID":"cd5f393a-8e7d-4fec-ba9c-362d5579a518","Type":"ContainerStarted","Data":"ab56534be10c77c8bfddbec0f316bacfd80299fad5163550786fd0b2ade6a8ad"} Apr 22 14:14:49.996999 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.996981 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" event={"ID":"17be8cc2ca560f8da5581c8d587a8582","Type":"ContainerStarted","Data":"a964830720cf7a894b37e10c7b562dd186e34b0898f6beac705a3e1f00430e35"} Apr 22 14:14:49.997926 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.997899 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-kg26l" event={"ID":"67d6b18b-35ca-4116-935f-5107cc660741","Type":"ContainerStarted","Data":"430190a23842cb6b4d20224c7fd9a02a2714a026da84e10f6e36daf0a7ba94ea"} Apr 22 14:14:49.998896 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.998872 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerStarted","Data":"a55e5b6ed4f633aae8a7620b48637270cb03a810e8bd2bd7ea42087b8041441a"} Apr 22 14:14:49.999977 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:49.999953 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-f42dv" event={"ID":"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a","Type":"ContainerStarted","Data":"71e7b6a66d7daf60efa3c067e4b33fcf0aac15affa1e8130d90a376a6f50476a"} Apr 22 14:14:50.009322 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:50.009286 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-136-18.ec2.internal" podStartSLOduration=2.009277417 podStartE2EDuration="2.009277417s" podCreationTimestamp="2026-04-22 14:14:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:14:50.009107961 +0000 UTC m=+3.476985690" watchObservedRunningTime="2026-04-22 14:14:50.009277417 +0000 UTC m=+3.477155146" Apr 22 14:14:50.608210 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:50.608117 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:50.608396 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:50.608281 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:50.608396 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:50.608343 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:14:52.608326216 +0000 UTC m=+6.076203928 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:50.709197 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:50.709163 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:50.709403 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:50.709389 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 14:14:50.709472 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:50.709413 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 14:14:50.709472 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:50.709426 2574 projected.go:194] Error preparing data for projected volume kube-api-access-85plq for pod openshift-network-diagnostics/network-check-target-68xhd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:50.709580 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:50.709486 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq podName:f67ee112-d296-44c7-9c63-b48da0c3bc44 nodeName:}" failed. No retries permitted until 2026-04-22 14:14:52.709468964 +0000 UTC m=+6.177346678 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-85plq" (UniqueName: "kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq") pod "network-check-target-68xhd" (UID: "f67ee112-d296-44c7-9c63-b48da0c3bc44") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:50.983742 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:50.983712 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:50.984160 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:50.983831 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:14:51.024772 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:51.024737 2574 generic.go:358] "Generic (PLEG): container finished" podID="e45c97e69339b1e77c7fd7ba3167751b" containerID="94ec06e9cc2e30932beb1e8a3d5910b7bda7c3abfe7baa327b1eb28b9d0a2345" exitCode=0 Apr 22 14:14:51.025261 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:51.025239 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" event={"ID":"e45c97e69339b1e77c7fd7ba3167751b","Type":"ContainerDied","Data":"94ec06e9cc2e30932beb1e8a3d5910b7bda7c3abfe7baa327b1eb28b9d0a2345"} Apr 22 14:14:51.983149 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:51.983114 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:51.983313 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:51.983260 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:14:52.035493 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:52.034839 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" event={"ID":"e45c97e69339b1e77c7fd7ba3167751b","Type":"ContainerStarted","Data":"cec7b9cdcc943f7851f6e3b4258bd7f0fbd5cbadefa8397f75a5194b45e5ca0c"} Apr 22 14:14:52.625536 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:52.625467 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:52.625745 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:52.625614 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:52.625745 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:52.625706 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:14:56.625687648 +0000 UTC m=+10.093565370 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:52.726148 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:52.726113 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:52.726317 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:52.726252 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 14:14:52.726317 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:52.726270 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 14:14:52.726317 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:52.726279 2574 projected.go:194] Error preparing data for projected volume kube-api-access-85plq for pod openshift-network-diagnostics/network-check-target-68xhd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:52.726493 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:52.726322 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq podName:f67ee112-d296-44c7-9c63-b48da0c3bc44 nodeName:}" failed. No retries permitted until 2026-04-22 14:14:56.726309471 +0000 UTC m=+10.194187179 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-85plq" (UniqueName: "kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq") pod "network-check-target-68xhd" (UID: "f67ee112-d296-44c7-9c63-b48da0c3bc44") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:52.985735 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:52.985260 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:52.985735 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:52.985380 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:14:53.983341 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:53.983308 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:53.983782 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:53.983450 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:14:54.983977 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:54.983947 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:54.984353 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:54.984043 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:14:55.983368 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:55.983328 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:55.983536 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:55.983511 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:14:56.660764 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:56.660718 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:56.661269 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:56.660905 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:56.661269 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:56.660959 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:04.660942218 +0000 UTC m=+18.128819928 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:14:56.761616 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:56.761578 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:56.762218 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:56.761787 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 14:14:56.762218 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:56.761813 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 14:14:56.762218 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:56.761826 2574 projected.go:194] Error preparing data for projected volume kube-api-access-85plq for pod openshift-network-diagnostics/network-check-target-68xhd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:56.762218 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:56.761882 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq podName:f67ee112-d296-44c7-9c63-b48da0c3bc44 nodeName:}" failed. No retries permitted until 2026-04-22 14:15:04.761864868 +0000 UTC m=+18.229742580 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-85plq" (UniqueName: "kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq") pod "network-check-target-68xhd" (UID: "f67ee112-d296-44c7-9c63-b48da0c3bc44") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:14:56.985336 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:56.984405 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:56.985336 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:56.984517 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:14:57.116146 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.116098 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-136-18.ec2.internal" podStartSLOduration=9.11608544 podStartE2EDuration="9.11608544s" podCreationTimestamp="2026-04-22 14:14:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:14:52.047825741 +0000 UTC m=+5.515703472" watchObservedRunningTime="2026-04-22 14:14:57.11608544 +0000 UTC m=+10.583963170" Apr 22 14:14:57.116649 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.116632 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-zfm4b"] Apr 22 14:14:57.122703 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.122260 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.122703 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:57.122353 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:14:57.164400 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.164373 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/3d7a8b58-1717-408a-9901-09f287c3731b-dbus\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.164537 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.164411 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.164537 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.164440 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/3d7a8b58-1717-408a-9901-09f287c3731b-kubelet-config\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.265223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.265147 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/3d7a8b58-1717-408a-9901-09f287c3731b-dbus\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.265223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.265188 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.265223 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.265220 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/3d7a8b58-1717-408a-9901-09f287c3731b-kubelet-config\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.265440 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.265316 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/3d7a8b58-1717-408a-9901-09f287c3731b-kubelet-config\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.265440 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.265438 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/3d7a8b58-1717-408a-9901-09f287c3731b-dbus\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.265525 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:57.265510 2574 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 14:14:57.265621 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:57.265571 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret podName:3d7a8b58-1717-408a-9901-09f287c3731b nodeName:}" failed. No retries permitted until 2026-04-22 14:14:57.765553232 +0000 UTC m=+11.233430944 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret") pod "global-pull-secret-syncer-zfm4b" (UID: "3d7a8b58-1717-408a-9901-09f287c3731b") : object "kube-system"/"original-pull-secret" not registered Apr 22 14:14:57.770852 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.770806 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:57.771285 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:57.770975 2574 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 14:14:57.771285 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:57.771030 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret podName:3d7a8b58-1717-408a-9901-09f287c3731b nodeName:}" failed. No retries permitted until 2026-04-22 14:14:58.771012814 +0000 UTC m=+12.238890527 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret") pod "global-pull-secret-syncer-zfm4b" (UID: "3d7a8b58-1717-408a-9901-09f287c3731b") : object "kube-system"/"original-pull-secret" not registered Apr 22 14:14:57.984022 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:57.983987 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:57.984184 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:57.984122 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:14:58.778441 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:58.778411 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:58.778889 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:58.778533 2574 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 14:14:58.778889 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:58.778592 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret podName:3d7a8b58-1717-408a-9901-09f287c3731b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:00.778580562 +0000 UTC m=+14.246458270 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret") pod "global-pull-secret-syncer-zfm4b" (UID: "3d7a8b58-1717-408a-9901-09f287c3731b") : object "kube-system"/"original-pull-secret" not registered Apr 22 14:14:58.984041 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:58.983967 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:14:58.984041 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:58.983998 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:14:58.984239 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:58.984094 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:14:58.984239 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:58.984182 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:14:59.983336 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:14:59.983299 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:14:59.983784 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:14:59.983445 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:00.793899 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:00.793864 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:00.794064 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:00.793996 2574 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 14:15:00.794064 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:00.794044 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret podName:3d7a8b58-1717-408a-9901-09f287c3731b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:04.794031863 +0000 UTC m=+18.261909571 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret") pod "global-pull-secret-syncer-zfm4b" (UID: "3d7a8b58-1717-408a-9901-09f287c3731b") : object "kube-system"/"original-pull-secret" not registered Apr 22 14:15:00.983096 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:00.983065 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:00.983279 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:00.983176 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:00.983279 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:00.983223 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:00.983376 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:00.983326 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:01.983252 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:01.983225 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:01.983421 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:01.983356 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:02.983571 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:02.983540 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:02.984025 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:02.983589 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:02.984025 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:02.983688 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:02.984025 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:02.983820 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:03.983348 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:03.983317 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:03.983591 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:03.983428 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:04.723015 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:04.722979 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:04.723369 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.723114 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:15:04.723369 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.723175 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:20.723157656 +0000 UTC m=+34.191035387 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:15:04.824090 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:04.824057 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:04.824237 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:04.824107 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:04.824237 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.824227 2574 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 14:15:04.824319 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.824258 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 14:15:04.824319 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.824276 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret podName:3d7a8b58-1717-408a-9901-09f287c3731b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:12.824263868 +0000 UTC m=+26.292141576 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret") pod "global-pull-secret-syncer-zfm4b" (UID: "3d7a8b58-1717-408a-9901-09f287c3731b") : object "kube-system"/"original-pull-secret" not registered Apr 22 14:15:04.824319 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.824284 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 14:15:04.824319 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.824298 2574 projected.go:194] Error preparing data for projected volume kube-api-access-85plq for pod openshift-network-diagnostics/network-check-target-68xhd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:15:04.824476 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.824350 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq podName:f67ee112-d296-44c7-9c63-b48da0c3bc44 nodeName:}" failed. No retries permitted until 2026-04-22 14:15:20.824334715 +0000 UTC m=+34.292212429 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-85plq" (UniqueName: "kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq") pod "network-check-target-68xhd" (UID: "f67ee112-d296-44c7-9c63-b48da0c3bc44") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:15:04.983900 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:04.983830 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:04.983900 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:04.983872 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:04.984094 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.983943 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:04.984094 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:04.984059 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:05.983200 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:05.983171 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:05.983556 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:05.983270 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:06.984612 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:06.984446 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:06.985238 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:06.984521 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:06.985238 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:06.984716 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:06.985238 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:06.984782 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:07.058724 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.058667 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-kg26l" event={"ID":"67d6b18b-35ca-4116-935f-5107cc660741","Type":"ContainerStarted","Data":"017f44cf36517f5c685ce07aa59081a7a1a6f2538e71f567c1c1494e5ecea01d"} Apr 22 14:15:07.060217 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.060191 2574 generic.go:358] "Generic (PLEG): container finished" podID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" containerID="a1634c1210a8908d38d2a2b5292b079f35fb7693066f3ae43dcd61ebee0ec874" exitCode=0 Apr 22 14:15:07.060315 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.060260 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerDied","Data":"a1634c1210a8908d38d2a2b5292b079f35fb7693066f3ae43dcd61ebee0ec874"} Apr 22 14:15:07.061607 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.061581 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-f42dv" event={"ID":"4cf8ecc7-e6d0-4d13-9d56-8e329378e84a","Type":"ContainerStarted","Data":"64ffbf5fb5eb0e9d571dd92252b6324e1ad1d99fbfe903e47e93a30bbf01a10f"} Apr 22 14:15:07.063227 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.063206 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hwrgf" event={"ID":"4ec55872-0394-47c2-9108-d21ab221b5f8","Type":"ContainerStarted","Data":"e1522c34482430fbae1f1c3d258abfe6ac379d11c718c05f5fb2ed88bbdc3911"} Apr 22 14:15:07.065085 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.065067 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" event={"ID":"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7","Type":"ContainerStarted","Data":"a70216e4b5612dbc2e20415ee15b1c223a96b88fc298697c909e10f8bf7c8ffc"} Apr 22 14:15:07.067517 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.067497 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:15:07.067929 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.067812 2574 generic.go:358] "Generic (PLEG): container finished" podID="3cff0f27-e53b-4843-a2ed-7e885a500c4a" containerID="79e6c6ea9dcbf8d77f2b441573c28ee6c000113efe6976a0a8ee1f66ed413274" exitCode=1 Apr 22 14:15:07.067929 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.067866 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"a4860fddf24f15d700609c73cf3b2ee158103e69fd1bc5fb18423021d6a3dd56"} Apr 22 14:15:07.067929 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.067884 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"c489c08f98cbfae99c40d75c59b5b70fab89b3b40dca6f3fe4928660575a4067"} Apr 22 14:15:07.067929 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.067899 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerDied","Data":"79e6c6ea9dcbf8d77f2b441573c28ee6c000113efe6976a0a8ee1f66ed413274"} Apr 22 14:15:07.067929 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.067913 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"d5c53e8cae1bf32dab9c4d36f0cd34c397e465f3c305b377c5bee7767a58a754"} Apr 22 14:15:07.069188 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.069169 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dx5kd" event={"ID":"3688a996-ab5a-4d60-b8e0-19b8a75bae48","Type":"ContainerStarted","Data":"e2caa0a7cfd6ba9408b3a1b08c5a90f50a088e83580b706c744014c0b49969ad"} Apr 22 14:15:07.070604 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.070583 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-d5pcf" event={"ID":"cd5f393a-8e7d-4fec-ba9c-362d5579a518","Type":"ContainerStarted","Data":"5115852acbe1b442492c4c74f38adc5ef71ad8391e7464d249ece625da40132a"} Apr 22 14:15:07.071269 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.071223 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-kg26l" podStartSLOduration=3.502048654 podStartE2EDuration="20.071209125s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.723038884 +0000 UTC m=+3.190916594" lastFinishedPulling="2026-04-22 14:15:06.29219935 +0000 UTC m=+19.760077065" observedRunningTime="2026-04-22 14:15:07.070723083 +0000 UTC m=+20.538600812" watchObservedRunningTime="2026-04-22 14:15:07.071209125 +0000 UTC m=+20.539086856" Apr 22 14:15:07.083340 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.083308 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-hwrgf" podStartSLOduration=3.464735711 podStartE2EDuration="20.083298237s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.717868774 +0000 UTC m=+3.185746482" lastFinishedPulling="2026-04-22 14:15:06.336431279 +0000 UTC m=+19.804309008" observedRunningTime="2026-04-22 14:15:07.083179613 +0000 UTC m=+20.551057344" watchObservedRunningTime="2026-04-22 14:15:07.083298237 +0000 UTC m=+20.551175967" Apr 22 14:15:07.096539 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.096506 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-d5pcf" podStartSLOduration=3.519298826 podStartE2EDuration="20.096496619s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.713748646 +0000 UTC m=+3.181626354" lastFinishedPulling="2026-04-22 14:15:06.290946439 +0000 UTC m=+19.758824147" observedRunningTime="2026-04-22 14:15:07.096294021 +0000 UTC m=+20.564171786" watchObservedRunningTime="2026-04-22 14:15:07.096496619 +0000 UTC m=+20.564374347" Apr 22 14:15:07.120623 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.120588 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dx5kd" podStartSLOduration=11.19795681 podStartE2EDuration="20.120577894s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.722504208 +0000 UTC m=+3.190381917" lastFinishedPulling="2026-04-22 14:14:58.645125278 +0000 UTC m=+12.113003001" observedRunningTime="2026-04-22 14:15:07.119935774 +0000 UTC m=+20.587813507" watchObservedRunningTime="2026-04-22 14:15:07.120577894 +0000 UTC m=+20.588455624" Apr 22 14:15:07.134887 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.134666 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-f42dv" podStartSLOduration=3.565341296 podStartE2EDuration="20.134652359s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.721235463 +0000 UTC m=+3.189113184" lastFinishedPulling="2026-04-22 14:15:06.290546524 +0000 UTC m=+19.758424247" observedRunningTime="2026-04-22 14:15:07.134228195 +0000 UTC m=+20.602105928" watchObservedRunningTime="2026-04-22 14:15:07.134652359 +0000 UTC m=+20.602530089" Apr 22 14:15:07.748098 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.748068 2574 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 22 14:15:07.952151 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.952046 2574 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-22T14:15:07.74809481Z","UUID":"3688e870-79fc-4cf5-ab1d-e37aa902d10b","Handler":null,"Name":"","Endpoint":""} Apr 22 14:15:07.954269 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.954244 2574 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 22 14:15:07.954386 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.954276 2574 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 22 14:15:07.983733 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:07.983705 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:07.983853 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:07.983828 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:08.074361 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:08.074326 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-9dhqr" event={"ID":"6a01b9d7-2c6c-4443-a988-d06252bd1a67","Type":"ContainerStarted","Data":"58ad55f0ccb0d7e1e9c520843ea0e3e9cbbbbc69f5e3623995acc427ad495a8c"} Apr 22 14:15:08.076304 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:08.076271 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" event={"ID":"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7","Type":"ContainerStarted","Data":"2d7de41e66ae7a017293ec10cbc4b82fbd8c0a5c4bbdfbc4c9b1c31863c9fa45"} Apr 22 14:15:08.079008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:08.078990 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:15:08.079473 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:08.079438 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"9e1222fb00e36abd1737725cc3238d86232138592ee4f0aabbe1c2841e8a86a3"} Apr 22 14:15:08.079473 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:08.079465 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"4f2e401a35fe3dfaedd4a6581566ea94593638da8beab9007598f6c2e98950f1"} Apr 22 14:15:08.983625 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:08.983587 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:08.983826 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:08.983586 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:08.983826 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:08.983716 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:08.983826 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:08.983799 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:09.400639 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:09.400613 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:15:09.401168 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:09.401147 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:15:09.429239 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:09.429192 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-9dhqr" podStartSLOduration=5.851382442 podStartE2EDuration="22.429177283s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.712798695 +0000 UTC m=+3.180676436" lastFinishedPulling="2026-04-22 14:15:06.290593564 +0000 UTC m=+19.758471277" observedRunningTime="2026-04-22 14:15:08.087830804 +0000 UTC m=+21.555708533" watchObservedRunningTime="2026-04-22 14:15:09.429177283 +0000 UTC m=+22.897055014" Apr 22 14:15:09.983381 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:09.983207 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:09.983565 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:09.983469 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:10.084537 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:10.084490 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" event={"ID":"9ca5ec93-ab8a-4af8-a32b-6ef0c190fcf7","Type":"ContainerStarted","Data":"b51aa5733677ad848b6d036514f14d951aa6b900a489fb6ac0cf478591418fbd"} Apr 22 14:15:10.088371 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:10.088341 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:15:10.089364 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:10.089340 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"a65a99113112f05e9e9eb08325962ead07357106a45ed1e3d7d94acad6fe865c"} Apr 22 14:15:10.104371 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:10.104329 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-nnps7" podStartSLOduration=3.7995053050000003 podStartE2EDuration="23.104318314s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.725947419 +0000 UTC m=+3.193825127" lastFinishedPulling="2026-04-22 14:15:09.030760428 +0000 UTC m=+22.498638136" observedRunningTime="2026-04-22 14:15:10.103775376 +0000 UTC m=+23.571653119" watchObservedRunningTime="2026-04-22 14:15:10.104318314 +0000 UTC m=+23.572196066" Apr 22 14:15:10.983464 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:10.983424 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:10.983934 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:10.983556 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:10.983934 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:10.983612 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:10.983934 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:10.983747 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:11.983414 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:11.983262 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:11.983576 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:11.983510 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:12.094112 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.094089 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:15:12.094457 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.094436 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"7e6c4e3195d95cf529c3c3c044b934958000c5ddced83df00dc8f7cf767e7bf1"} Apr 22 14:15:12.094728 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.094709 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:15:12.094944 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.094912 2574 scope.go:117] "RemoveContainer" containerID="79e6c6ea9dcbf8d77f2b441573c28ee6c000113efe6976a0a8ee1f66ed413274" Apr 22 14:15:12.096134 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.096116 2574 generic.go:358] "Generic (PLEG): container finished" podID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" containerID="3d89437c0c80b11244812669fe735476cfa4a89aa98b60126559d603de5d4f35" exitCode=0 Apr 22 14:15:12.096218 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.096149 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerDied","Data":"3d89437c0c80b11244812669fe735476cfa4a89aa98b60126559d603de5d4f35"} Apr 22 14:15:12.110113 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.110097 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:15:12.889693 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.889649 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:12.889823 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:12.889778 2574 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 14:15:12.889869 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:12.889840 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret podName:3d7a8b58-1717-408a-9901-09f287c3731b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:28.889826032 +0000 UTC m=+42.357703746 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret") pod "global-pull-secret-syncer-zfm4b" (UID: "3d7a8b58-1717-408a-9901-09f287c3731b") : object "kube-system"/"original-pull-secret" not registered Apr 22 14:15:12.983475 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.983413 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:12.983475 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:12.983428 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:12.983854 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:12.983499 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:12.983854 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:12.983633 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:13.001514 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:13.001472 2574 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = unable to pull image or OCI artifact: pull image err: copying system image from manifest list: reading blob sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803: Digest did not match, expected sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855; artifact err: provided artifact is a container image" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91034a2e8fa729a060ad18831a3c6e5de5d2b7b3de437b198ddc24fcb724dcf6" Apr 22 14:15:13.001707 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:13.001656 2574 kuberuntime_manager.go:1358] "Unhandled Error" err="init container &Container{Name:bond-cni-plugin,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91034a2e8fa729a060ad18831a3c6e5de5d2b7b3de437b198ddc24fcb724dcf6,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/bondcni/rhel8/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/bondcni/rhel9/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/bondcni/rhel9/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z6hbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-25xzb_openshift-multus(cc6913ef-cc9f-4a41-971f-4e5dda9212fd): ErrImagePull: unable to pull image or OCI artifact: pull image err: copying system image from manifest list: reading blob sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803: Digest did not match, expected sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855; artifact err: provided artifact is a container image" logger="UnhandledError" Apr 22 14:15:13.002825 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:13.002802 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bond-cni-plugin\" with ErrImagePull: \"unable to pull image or OCI artifact: pull image err: copying system image from manifest list: reading blob sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803: Digest did not match, expected sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855; artifact err: provided artifact is a container image\"" pod="openshift-multus/multus-additional-cni-plugins-25xzb" podUID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" Apr 22 14:15:13.019866 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.019846 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:15:13.019959 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.019945 2574 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 22 14:15:13.020423 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.020400 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-f42dv" Apr 22 14:15:13.101533 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.101509 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:15:13.101927 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.101903 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" event={"ID":"3cff0f27-e53b-4843-a2ed-7e885a500c4a","Type":"ContainerStarted","Data":"a119140a6db30c5a87e3b002939bb71ffbe74458ba7bac9b8bc482340bd7943b"} Apr 22 14:15:13.102027 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.101950 2574 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 22 14:15:13.102185 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.102149 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:15:13.103270 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:13.103188 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bond-cni-plugin\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91034a2e8fa729a060ad18831a3c6e5de5d2b7b3de437b198ddc24fcb724dcf6\\\": ErrImagePull: unable to pull image or OCI artifact: pull image err: copying system image from manifest list: reading blob sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803: Digest did not match, expected sha256:fa95077c3721855d828504f9b4991658ad4537426b29860688fe0b88d808d803, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855; artifact err: provided artifact is a container image\"" pod="openshift-multus/multus-additional-cni-plugins-25xzb" podUID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" Apr 22 14:15:13.116177 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.116160 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:15:13.132095 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.132058 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" podStartSLOduration=9.316327936 podStartE2EDuration="26.13204701s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.726235257 +0000 UTC m=+3.194112966" lastFinishedPulling="2026-04-22 14:15:06.541954319 +0000 UTC m=+20.009832040" observedRunningTime="2026-04-22 14:15:13.13153776 +0000 UTC m=+26.599415491" watchObservedRunningTime="2026-04-22 14:15:13.13204701 +0000 UTC m=+26.599924740" Apr 22 14:15:13.518027 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.517997 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-zfm4b"] Apr 22 14:15:13.518181 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.518096 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:13.518214 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:13.518180 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:13.521224 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.521197 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-68xhd"] Apr 22 14:15:13.521350 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.521287 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:13.521407 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:13.521388 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:13.521755 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.521733 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p65qg"] Apr 22 14:15:13.521853 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:13.521841 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:13.521967 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:13.521948 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:14.103337 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:14.103314 2574 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 22 14:15:14.983488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:14.983457 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:14.983488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:14.983478 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:14.983739 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:14.983461 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:14.983739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:14.983565 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:14.983739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:14.983620 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:14.983739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:14.983696 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:15.104809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:15.104785 2574 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 22 14:15:16.983827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:16.983800 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:16.984197 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:16.983899 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:16.984197 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:16.983951 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:16.984197 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:16.984014 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:16.984197 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:16.984039 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:16.984197 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:16.984124 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:18.875956 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:18.875919 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:15:18.876339 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:18.876138 2574 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 22 14:15:18.894430 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:18.894278 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpjz8" Apr 22 14:15:18.983480 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:18.983441 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:18.983617 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:18.983547 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:18.983617 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:18.983556 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-zfm4b" podUID="3d7a8b58-1717-408a-9901-09f287c3731b" Apr 22 14:15:18.983734 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:18.983632 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-68xhd" podUID="f67ee112-d296-44c7-9c63-b48da0c3bc44" Apr 22 14:15:18.983734 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:18.983692 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:18.983795 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:18.983752 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:15:19.430842 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.430815 2574 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-136-18.ec2.internal" event="NodeReady" Apr 22 14:15:19.431003 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.430954 2574 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 22 14:15:19.482607 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.482580 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-7bfcd86c87-jl5cs"] Apr 22 14:15:19.514842 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.514817 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-7bfcd86c87-jl5cs"] Apr 22 14:15:19.514842 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.514842 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-4kpkf"] Apr 22 14:15:19.515072 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.514989 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.517874 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.517818 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 22 14:15:19.518001 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.517977 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 22 14:15:19.518049 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.517981 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-bgh6b\"" Apr 22 14:15:19.518049 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.518024 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 22 14:15:19.539411 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.539386 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 22 14:15:19.539411 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.539401 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b6brs"] Apr 22 14:15:19.539606 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.539588 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.542883 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.542863 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 22 14:15:19.542976 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.542959 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-hhqr7\"" Apr 22 14:15:19.543076 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.543068 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 22 14:15:19.554467 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.554450 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4kpkf"] Apr 22 14:15:19.554467 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.554470 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b6brs"] Apr 22 14:15:19.554596 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.554539 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:19.557022 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.556997 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 22 14:15:19.557109 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.557051 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 22 14:15:19.557294 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.557279 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-qh95w\"" Apr 22 14:15:19.557483 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.557469 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 22 14:15:19.641588 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641557 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-certificates\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.641795 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641610 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.641795 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641714 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:19.641795 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641744 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-ca-trust-extracted\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.641795 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641760 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-trusted-ca\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.641795 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641779 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dck6x\" (UniqueName: \"kubernetes.io/projected/daaf870a-47fd-4fba-8f87-375c235c1a4c-kube-api-access-dck6x\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:19.641979 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641861 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-image-registry-private-configuration\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.641979 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641890 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-installation-pull-secrets\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.641979 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641907 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-config-volume\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.641979 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641931 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnsj7\" (UniqueName: \"kubernetes.io/projected/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-kube-api-access-cnsj7\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.642150 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.641991 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-tmp-dir\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.642150 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.642031 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-bound-sa-token\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.642150 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.642060 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.642150 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.642088 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khslj\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-kube-api-access-khslj\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.742495 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742411 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:19.742495 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742447 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-ca-trust-extracted\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.742495 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742468 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-trusted-ca\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.742495 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742488 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dck6x\" (UniqueName: \"kubernetes.io/projected/daaf870a-47fd-4fba-8f87-375c235c1a4c-kube-api-access-dck6x\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742513 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-image-registry-private-configuration\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742538 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-installation-pull-secrets\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742554 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-config-volume\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:19.742565 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:19.742643 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:15:20.242623363 +0000 UTC m=+33.710501083 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742568 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cnsj7\" (UniqueName: \"kubernetes.io/projected/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-kube-api-access-cnsj7\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742767 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-tmp-dir\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742801 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-bound-sa-token\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.742827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742831 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742861 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-khslj\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-kube-api-access-khslj\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742905 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-certificates\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.742937 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:19.742986 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:19.743002 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:19.743032 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:19.743058 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:15:20.243042245 +0000 UTC m=+33.710919967 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:19.743084 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:20.243070222 +0000 UTC m=+33.710947930 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.743123 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-tmp-dir\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.743243 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.743211 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-config-volume\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.743706 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.743393 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-ca-trust-extracted\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.743706 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.743545 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-trusted-ca\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.743706 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.743525 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-certificates\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.746638 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.746618 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-installation-pull-secrets\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.746739 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.746625 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-image-registry-private-configuration\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.752348 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.752324 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnsj7\" (UniqueName: \"kubernetes.io/projected/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-kube-api-access-cnsj7\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:19.752835 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.752813 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-bound-sa-token\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:19.752950 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.752930 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dck6x\" (UniqueName: \"kubernetes.io/projected/daaf870a-47fd-4fba-8f87-375c235c1a4c-kube-api-access-dck6x\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:19.753134 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:19.753115 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-khslj\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-kube-api-access-khslj\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:20.247932 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.247898 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.247967 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.248022 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.248051 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.248103 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.248117 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:15:21.248098061 +0000 UTC m=+34.715975783 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.248133 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:21.248123941 +0000 UTC m=+34.716001649 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.248137 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.248161 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:15:20.248458 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.248231 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:15:21.24821313 +0000 UTC m=+34.716090846 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:15:20.751088 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.751059 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:20.751324 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.751232 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:15:20.751324 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.751302 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:52.751285632 +0000 UTC m=+66.219163341 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 14:15:20.852403 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.852363 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:20.852593 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.852538 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 14:15:20.852593 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.852561 2574 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 14:15:20.852593 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.852571 2574 projected.go:194] Error preparing data for projected volume kube-api-access-85plq for pod openshift-network-diagnostics/network-check-target-68xhd: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:15:20.852732 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:20.852621 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq podName:f67ee112-d296-44c7-9c63-b48da0c3bc44 nodeName:}" failed. No retries permitted until 2026-04-22 14:15:52.852606896 +0000 UTC m=+66.320484608 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-85plq" (UniqueName: "kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq") pod "network-check-target-68xhd" (UID: "f67ee112-d296-44c7-9c63-b48da0c3bc44") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 14:15:20.983648 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.983617 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:20.983916 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.983616 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:20.983916 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.983628 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:20.986376 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.986356 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 22 14:15:20.986503 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.986357 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-95blp\"" Apr 22 14:15:20.986503 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.986432 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 22 14:15:20.986939 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.986918 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 22 14:15:20.987018 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.986929 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-4nhsh\"" Apr 22 14:15:20.987018 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:20.986986 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 22 14:15:21.255465 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:21.255433 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:21.255492 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:21.255526 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:21.255607 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:21.255628 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:21.255629 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:21.255612 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:21.255697 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:23.255666032 +0000 UTC m=+36.723543740 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:21.255711 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:15:23.255704888 +0000 UTC m=+36.723582596 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:15:21.255980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:21.255733 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:15:23.255716451 +0000 UTC m=+36.723594173 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:15:23.270512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:23.270476 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:23.270512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:23.270528 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:23.270563 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:23.270636 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:23.270650 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:23.270661 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:23.270697 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:23.270731 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:15:27.270717562 +0000 UTC m=+40.738595270 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:23.270743 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:27.270738217 +0000 UTC m=+40.738615925 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:15:23.271075 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:23.270753 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:15:27.270747922 +0000 UTC m=+40.738625630 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:15:27.298226 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:27.298183 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:27.298249 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:27.298287 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:27.298321 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:27.298339 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:27.298389 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:15:35.298372546 +0000 UTC m=+48.766250259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:27.298398 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:27.298398 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:27.298443 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:15:35.298428509 +0000 UTC m=+48.766306217 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:15:27.298739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:27.298456 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:35.298450494 +0000 UTC m=+48.766328202 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:15:28.909320 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:28.909282 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:28.911596 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:28.911576 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/3d7a8b58-1717-408a-9901-09f287c3731b-original-pull-secret\") pod \"global-pull-secret-syncer-zfm4b\" (UID: \"3d7a8b58-1717-408a-9901-09f287c3731b\") " pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:29.093767 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:29.093744 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-zfm4b" Apr 22 14:15:29.133062 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:29.133028 2574 generic.go:358] "Generic (PLEG): container finished" podID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" containerID="dab79e80d54d626013ae1e0b004552eebe540f067d78eb9d731aecdfcbd249bd" exitCode=0 Apr 22 14:15:29.133177 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:29.133103 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerDied","Data":"dab79e80d54d626013ae1e0b004552eebe540f067d78eb9d731aecdfcbd249bd"} Apr 22 14:15:29.219119 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:29.219089 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-zfm4b"] Apr 22 14:15:29.223442 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:15:29.223416 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d7a8b58_1717_408a_9901_09f287c3731b.slice/crio-93bae7cddcc7117d7dc86dab0b1463883a2a9dcf7d37f8cbc834e3dd50cb4901 WatchSource:0}: Error finding container 93bae7cddcc7117d7dc86dab0b1463883a2a9dcf7d37f8cbc834e3dd50cb4901: Status 404 returned error can't find the container with id 93bae7cddcc7117d7dc86dab0b1463883a2a9dcf7d37f8cbc834e3dd50cb4901 Apr 22 14:15:30.136412 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:30.136374 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-zfm4b" event={"ID":"3d7a8b58-1717-408a-9901-09f287c3731b","Type":"ContainerStarted","Data":"93bae7cddcc7117d7dc86dab0b1463883a2a9dcf7d37f8cbc834e3dd50cb4901"} Apr 22 14:15:30.138768 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:30.138740 2574 generic.go:358] "Generic (PLEG): container finished" podID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" containerID="2d26b4f36db273d33e06751e90bb203454a82c0d68785b43a8cd62c3aaad200b" exitCode=0 Apr 22 14:15:30.138898 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:30.138780 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerDied","Data":"2d26b4f36db273d33e06751e90bb203454a82c0d68785b43a8cd62c3aaad200b"} Apr 22 14:15:35.360462 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:35.360418 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:35.360507 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:35.360545 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:35.360549 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:35.360569 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:35.360624 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:15:51.360609587 +0000 UTC m=+64.828487299 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:35.360658 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:35.360660 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:35.360730 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:15:51.360714827 +0000 UTC m=+64.828592538 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:15:35.360980 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:35.360749 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:15:51.360739117 +0000 UTC m=+64.828616834 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:15:37.153840 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:37.153781 2574 generic.go:358] "Generic (PLEG): container finished" podID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" containerID="9ba9c55b124e59c0a2e3e979aa98b54e731285309e62d5bf696a4ae688ac1328" exitCode=0 Apr 22 14:15:37.154236 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:37.153851 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerDied","Data":"9ba9c55b124e59c0a2e3e979aa98b54e731285309e62d5bf696a4ae688ac1328"} Apr 22 14:15:37.155045 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:37.155020 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-zfm4b" event={"ID":"3d7a8b58-1717-408a-9901-09f287c3731b","Type":"ContainerStarted","Data":"1eceb4a2aebf13d708909edf33f536106b8dc89909cef67ab732a3aca137ee42"} Apr 22 14:15:37.191132 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:37.190902 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-zfm4b" podStartSLOduration=33.085197017 podStartE2EDuration="40.190883065s" podCreationTimestamp="2026-04-22 14:14:57 +0000 UTC" firstStartedPulling="2026-04-22 14:15:29.225037038 +0000 UTC m=+42.692914746" lastFinishedPulling="2026-04-22 14:15:36.330723082 +0000 UTC m=+49.798600794" observedRunningTime="2026-04-22 14:15:37.190144217 +0000 UTC m=+50.658021972" watchObservedRunningTime="2026-04-22 14:15:37.190883065 +0000 UTC m=+50.658760797" Apr 22 14:15:38.159778 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:38.159749 2574 generic.go:358] "Generic (PLEG): container finished" podID="cc6913ef-cc9f-4a41-971f-4e5dda9212fd" containerID="25a02172170517a4c7f8a6fbe01cf5aecd9af1da06245cffd6c7670edb835acb" exitCode=0 Apr 22 14:15:38.160173 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:38.159824 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerDied","Data":"25a02172170517a4c7f8a6fbe01cf5aecd9af1da06245cffd6c7670edb835acb"} Apr 22 14:15:39.167645 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:39.167606 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-25xzb" event={"ID":"cc6913ef-cc9f-4a41-971f-4e5dda9212fd","Type":"ContainerStarted","Data":"c16c8429449700f2f1169d692b557405da629934d9e0fd59d3a4b2564027bff4"} Apr 22 14:15:39.196483 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:39.196438 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-25xzb" podStartSLOduration=5.56357792 podStartE2EDuration="52.196423356s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:14:49.718979452 +0000 UTC m=+3.186857174" lastFinishedPulling="2026-04-22 14:15:36.351824887 +0000 UTC m=+49.819702610" observedRunningTime="2026-04-22 14:15:39.194776999 +0000 UTC m=+52.662654728" watchObservedRunningTime="2026-04-22 14:15:39.196423356 +0000 UTC m=+52.664301086" Apr 22 14:15:51.372574 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:51.372531 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:51.372596 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:51.372618 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:51.372717 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:51.372722 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:51.372742 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:51.372753 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:51.372765 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:16:23.37275247 +0000 UTC m=+96.840630179 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:51.372806 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:16:23.372793223 +0000 UTC m=+96.840670931 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:15:51.372982 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:51.372825 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:16:23.372815509 +0000 UTC m=+96.840693221 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:15:52.782102 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:52.782054 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:15:52.784515 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:52.784499 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 22 14:15:52.792415 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:52.792399 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 22 14:15:52.792461 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:15:52.792453 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:16:56.792435451 +0000 UTC m=+130.260313160 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : secret "metrics-daemon-secret" not found Apr 22 14:15:52.882502 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:52.882467 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:52.884822 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:52.884806 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 22 14:15:52.895593 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:52.895576 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 22 14:15:52.906717 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:52.906693 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-85plq\" (UniqueName: \"kubernetes.io/projected/f67ee112-d296-44c7-9c63-b48da0c3bc44-kube-api-access-85plq\") pod \"network-check-target-68xhd\" (UID: \"f67ee112-d296-44c7-9c63-b48da0c3bc44\") " pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:53.101142 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:53.101072 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-95blp\"" Apr 22 14:15:53.109138 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:53.109116 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:53.219514 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:53.219481 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-68xhd"] Apr 22 14:15:53.222983 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:15:53.222957 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf67ee112_d296_44c7_9c63_b48da0c3bc44.slice/crio-f962bf26223c08273e95f08b2aac1fa29d59ed0dd18fec37dc797e0cfc672c29 WatchSource:0}: Error finding container f962bf26223c08273e95f08b2aac1fa29d59ed0dd18fec37dc797e0cfc672c29: Status 404 returned error can't find the container with id f962bf26223c08273e95f08b2aac1fa29d59ed0dd18fec37dc797e0cfc672c29 Apr 22 14:15:54.194965 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:54.194932 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-68xhd" event={"ID":"f67ee112-d296-44c7-9c63-b48da0c3bc44","Type":"ContainerStarted","Data":"f962bf26223c08273e95f08b2aac1fa29d59ed0dd18fec37dc797e0cfc672c29"} Apr 22 14:15:56.203564 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:56.203528 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-68xhd" event={"ID":"f67ee112-d296-44c7-9c63-b48da0c3bc44","Type":"ContainerStarted","Data":"5e2a9b8422402d7303fe5ec26bdb3ed4bdfe16948784a296fcd85e49ceda9ece"} Apr 22 14:15:56.203973 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:56.203709 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:15:56.219540 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:15:56.219507 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-68xhd" podStartSLOduration=66.760017103 podStartE2EDuration="1m9.219497893s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:15:53.22521264 +0000 UTC m=+66.693090348" lastFinishedPulling="2026-04-22 14:15:55.684693415 +0000 UTC m=+69.152571138" observedRunningTime="2026-04-22 14:15:56.219328421 +0000 UTC m=+69.687206163" watchObservedRunningTime="2026-04-22 14:15:56.219497893 +0000 UTC m=+69.687375623" Apr 22 14:16:23.393819 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:16:23.393775 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:16:23.393835 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:16:23.393877 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:23.393923 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:23.393979 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:23.393990 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:17:27.393974713 +0000 UTC m=+160.861852422 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:23.394006 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:23.394023 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:23.394038 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:17:27.394020852 +0000 UTC m=+160.861898566 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:16:23.394243 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:23.394068 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:17:27.394056422 +0000 UTC m=+160.861934130 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:16:27.207780 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:16:27.207747 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-68xhd" Apr 22 14:16:56.831327 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:16:56.831282 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:16:56.831837 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:56.831431 2574 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 22 14:16:56.831837 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:16:56.831505 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs podName:1eca98e2-015f-4ce7-bcd6-76878fa26d4b nodeName:}" failed. No retries permitted until 2026-04-22 14:18:58.831489004 +0000 UTC m=+252.299366712 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs") pod "network-metrics-daemon-p65qg" (UID: "1eca98e2-015f-4ce7-bcd6-76878fa26d4b") : secret "metrics-daemon-secret" not found Apr 22 14:17:18.061535 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.061499 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg"] Apr 22 14:17:18.064393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.064374 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:18.066613 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.066593 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Apr 22 14:17:18.067061 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.067045 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Apr 22 14:17:18.067135 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.067060 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Apr 22 14:17:18.067135 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.067106 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-nplht\"" Apr 22 14:17:18.075138 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.075120 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg"] Apr 22 14:17:18.084458 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.084439 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8lhg\" (UniqueName: \"kubernetes.io/projected/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-kube-api-access-v8lhg\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:18.084542 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.084466 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:18.169565 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.169523 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9"] Apr 22 14:17:18.172621 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.172600 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv"] Apr 22 14:17:18.172784 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.172768 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.175334 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.175314 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-operator-585dfdc468-zkfsf"] Apr 22 14:17:18.175456 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.175440 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:18.177926 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.177908 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.179325 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.179297 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Apr 22 14:17:18.179410 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.179351 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Apr 22 14:17:18.180746 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.180726 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-p4xrw\"" Apr 22 14:17:18.180842 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.180814 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Apr 22 14:17:18.180919 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.180898 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Apr 22 14:17:18.181411 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.181391 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Apr 22 14:17:18.183093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.183068 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"default-dockercfg-xnxkr\"" Apr 22 14:17:18.185538 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185517 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/727af150-86df-4722-a839-797b4cb1db69-tmp\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.185647 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185555 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36bf42f9-41dd-4d29-9f4c-2472e7df5572-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.185647 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185585 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727af150-86df-4722-a839-797b4cb1db69-serving-cert\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.185802 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185648 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36bf42f9-41dd-4d29-9f4c-2472e7df5572-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.185802 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185689 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7k9j\" (UniqueName: \"kubernetes.io/projected/36bf42f9-41dd-4d29-9f4c-2472e7df5572-kube-api-access-g7k9j\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.185802 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185733 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/727af150-86df-4722-a839-797b4cb1db69-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.185802 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185776 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6513f255-e88c-4c51-8d5b-17018a5c835d-nginx-conf\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:18.185802 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185801 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/727af150-86df-4722-a839-797b4cb1db69-snapshots\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.186051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185806 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Apr 22 14:17:18.186051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185828 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v8lhg\" (UniqueName: \"kubernetes.io/projected/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-kube-api-access-v8lhg\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:18.186051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185846 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:18.186051 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.185931 2574 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 22 14:17:18.186051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.185929 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/727af150-86df-4722-a839-797b4cb1db69-service-ca-bundle\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.186051 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.185988 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls podName:a3973b04-9fff-4f0a-bb01-3de9b19fba5a nodeName:}" failed. No retries permitted until 2026-04-22 14:17:18.685969751 +0000 UTC m=+152.153847460 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-z45rg" (UID: "a3973b04-9fff-4f0a-bb01-3de9b19fba5a") : secret "samples-operator-tls" not found Apr 22 14:17:18.186431 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.186054 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnbxn\" (UniqueName: \"kubernetes.io/projected/727af150-86df-4722-a839-797b4cb1db69-kube-api-access-mnbxn\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.186431 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.186112 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:18.186431 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.186199 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 22 14:17:18.186911 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.186890 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"service-ca-bundle\"" Apr 22 14:17:18.187014 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.186961 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"operator-dockercfg-wp27l\"" Apr 22 14:17:18.187014 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.186973 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 22 14:17:18.187691 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.187655 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"openshift-insights-serving-cert\"" Apr 22 14:17:18.191294 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.191277 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"trusted-ca-bundle\"" Apr 22 14:17:18.205986 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.205956 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-zkfsf"] Apr 22 14:17:18.207787 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.207765 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9"] Apr 22 14:17:18.208611 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.208593 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv"] Apr 22 14:17:18.215781 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.215762 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8lhg\" (UniqueName: \"kubernetes.io/projected/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-kube-api-access-v8lhg\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:18.283361 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.283326 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-pxmjh"] Apr 22 14:17:18.286296 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.286278 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn"] Apr 22 14:17:18.286509 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.286487 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.286968 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.286943 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/727af150-86df-4722-a839-797b4cb1db69-snapshots\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287089 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.286991 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-trusted-ca\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.287089 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287053 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/727af150-86df-4722-a839-797b4cb1db69-service-ca-bundle\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287089 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287083 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mnbxn\" (UniqueName: \"kubernetes.io/projected/727af150-86df-4722-a839-797b4cb1db69-kube-api-access-mnbxn\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287229 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287108 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:18.287229 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.287181 2574 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 22 14:17:18.287336 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.287236 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert podName:6513f255-e88c-4c51-8d5b-17018a5c835d nodeName:}" failed. No retries permitted until 2026-04-22 14:17:18.787218756 +0000 UTC m=+152.255096467 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-8fwqv" (UID: "6513f255-e88c-4c51-8d5b-17018a5c835d") : secret "networking-console-plugin-cert" not found Apr 22 14:17:18.287336 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287250 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/727af150-86df-4722-a839-797b4cb1db69-tmp\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287336 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287269 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqkbv\" (UniqueName: \"kubernetes.io/projected/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-kube-api-access-pqkbv\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.287336 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287284 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-serving-cert\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.287336 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287311 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36bf42f9-41dd-4d29-9f4c-2472e7df5572-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.287551 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287355 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727af150-86df-4722-a839-797b4cb1db69-serving-cert\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287551 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287416 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-config\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.287551 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287481 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36bf42f9-41dd-4d29-9f4c-2472e7df5572-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.287551 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287511 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g7k9j\" (UniqueName: \"kubernetes.io/projected/36bf42f9-41dd-4d29-9f4c-2472e7df5572-kube-api-access-g7k9j\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.287551 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287539 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/727af150-86df-4722-a839-797b4cb1db69-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287796 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287564 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6513f255-e88c-4c51-8d5b-17018a5c835d-nginx-conf\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:18.287796 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287618 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/727af150-86df-4722-a839-797b4cb1db69-service-ca-bundle\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287796 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287720 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/727af150-86df-4722-a839-797b4cb1db69-tmp\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.287934 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.287720 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/727af150-86df-4722-a839-797b4cb1db69-snapshots\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.288137 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.288115 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36bf42f9-41dd-4d29-9f4c-2472e7df5572-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.288746 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.288723 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6513f255-e88c-4c51-8d5b-17018a5c835d-nginx-conf\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:18.288946 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.288924 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/727af150-86df-4722-a839-797b4cb1db69-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.289530 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.289514 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" Apr 22 14:17:18.290013 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.289974 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36bf42f9-41dd-4d29-9f4c-2472e7df5572-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.290093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.289984 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727af150-86df-4722-a839-797b4cb1db69-serving-cert\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.292952 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.292935 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Apr 22 14:17:18.293019 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.292972 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Apr 22 14:17:18.293183 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.293167 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Apr 22 14:17:18.293280 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.293171 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Apr 22 14:17:18.293514 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.293495 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-26msb\"" Apr 22 14:17:18.293664 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.293651 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"network-diagnostics-dockercfg-j9gbl\"" Apr 22 14:17:18.299518 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.299494 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Apr 22 14:17:18.303568 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.303547 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-pxmjh"] Apr 22 14:17:18.306337 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.306313 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn"] Apr 22 14:17:18.314915 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.314859 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7k9j\" (UniqueName: \"kubernetes.io/projected/36bf42f9-41dd-4d29-9f4c-2472e7df5572-kube-api-access-g7k9j\") pod \"kube-storage-version-migrator-operator-6769c5d45-9lcw9\" (UID: \"36bf42f9-41dd-4d29-9f4c-2472e7df5572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.315048 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.315031 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnbxn\" (UniqueName: \"kubernetes.io/projected/727af150-86df-4722-a839-797b4cb1db69-kube-api-access-mnbxn\") pod \"insights-operator-585dfdc468-zkfsf\" (UID: \"727af150-86df-4722-a839-797b4cb1db69\") " pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.388426 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.388380 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pqkbv\" (UniqueName: \"kubernetes.io/projected/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-kube-api-access-pqkbv\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.388426 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.388423 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-serving-cert\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.388726 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.388455 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-config\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.388726 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.388497 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v85s\" (UniqueName: \"kubernetes.io/projected/1bdda416-04a6-4a10-a11e-ead137503dd2-kube-api-access-6v85s\") pod \"network-check-source-8894fc9bd-9m6vn\" (UID: \"1bdda416-04a6-4a10-a11e-ead137503dd2\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" Apr 22 14:17:18.388726 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.388531 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-trusted-ca\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.389185 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.389167 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-config\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.389276 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.389259 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-trusted-ca\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.390903 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.390887 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-serving-cert\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.397409 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.397392 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqkbv\" (UniqueName: \"kubernetes.io/projected/00643d7f-54c7-4ca9-8d93-e7cf0c7554c9-kube-api-access-pqkbv\") pod \"console-operator-9d4b6777b-pxmjh\" (UID: \"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9\") " pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.484495 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.484459 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" Apr 22 14:17:18.488940 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.488915 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6v85s\" (UniqueName: \"kubernetes.io/projected/1bdda416-04a6-4a10-a11e-ead137503dd2-kube-api-access-6v85s\") pod \"network-check-source-8894fc9bd-9m6vn\" (UID: \"1bdda416-04a6-4a10-a11e-ead137503dd2\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" Apr 22 14:17:18.495102 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.495065 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-zkfsf" Apr 22 14:17:18.497227 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.497191 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v85s\" (UniqueName: \"kubernetes.io/projected/1bdda416-04a6-4a10-a11e-ead137503dd2-kube-api-access-6v85s\") pod \"network-check-source-8894fc9bd-9m6vn\" (UID: \"1bdda416-04a6-4a10-a11e-ead137503dd2\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" Apr 22 14:17:18.606429 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.606405 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:18.610856 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.610821 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9"] Apr 22 14:17:18.612333 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.611862 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" Apr 22 14:17:18.615358 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:18.615331 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36bf42f9_41dd_4d29_9f4c_2472e7df5572.slice/crio-a4bdc0e74d88b1d79e872e229229b0f8311e8f34b5f41b8d1e8a4e5682e7fa18 WatchSource:0}: Error finding container a4bdc0e74d88b1d79e872e229229b0f8311e8f34b5f41b8d1e8a4e5682e7fa18: Status 404 returned error can't find the container with id a4bdc0e74d88b1d79e872e229229b0f8311e8f34b5f41b8d1e8a4e5682e7fa18 Apr 22 14:17:18.624894 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.624865 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-zkfsf"] Apr 22 14:17:18.628990 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:18.628960 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod727af150_86df_4722_a839_797b4cb1db69.slice/crio-f0ed4f7b3da55373dc6275ecd6e365066a721d1d1ce79cccbf429ffdd7baed7c WatchSource:0}: Error finding container f0ed4f7b3da55373dc6275ecd6e365066a721d1d1ce79cccbf429ffdd7baed7c: Status 404 returned error can't find the container with id f0ed4f7b3da55373dc6275ecd6e365066a721d1d1ce79cccbf429ffdd7baed7c Apr 22 14:17:18.691415 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.691391 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:18.691571 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.691549 2574 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 22 14:17:18.691642 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.691632 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls podName:a3973b04-9fff-4f0a-bb01-3de9b19fba5a nodeName:}" failed. No retries permitted until 2026-04-22 14:17:19.691611258 +0000 UTC m=+153.159488979 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-z45rg" (UID: "a3973b04-9fff-4f0a-bb01-3de9b19fba5a") : secret "samples-operator-tls" not found Apr 22 14:17:18.728268 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.728238 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-pxmjh"] Apr 22 14:17:18.731557 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:18.731534 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00643d7f_54c7_4ca9_8d93_e7cf0c7554c9.slice/crio-c0c1bdbea8a22c79e0c75fd7a5b3c4843c4371c9f954b86c2449d49e9f7143e0 WatchSource:0}: Error finding container c0c1bdbea8a22c79e0c75fd7a5b3c4843c4371c9f954b86c2449d49e9f7143e0: Status 404 returned error can't find the container with id c0c1bdbea8a22c79e0c75fd7a5b3c4843c4371c9f954b86c2449d49e9f7143e0 Apr 22 14:17:18.742946 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.742906 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn"] Apr 22 14:17:18.748262 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:18.748230 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bdda416_04a6_4a10_a11e_ead137503dd2.slice/crio-9a39a214a9c8ad601a004e62f19a8729d944420ecf6ac3bef2e8a18e5369880c WatchSource:0}: Error finding container 9a39a214a9c8ad601a004e62f19a8729d944420ecf6ac3bef2e8a18e5369880c: Status 404 returned error can't find the container with id 9a39a214a9c8ad601a004e62f19a8729d944420ecf6ac3bef2e8a18e5369880c Apr 22 14:17:18.792556 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:18.792530 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:18.792739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.792635 2574 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 22 14:17:18.792739 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:18.792701 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert podName:6513f255-e88c-4c51-8d5b-17018a5c835d nodeName:}" failed. No retries permitted until 2026-04-22 14:17:19.792687225 +0000 UTC m=+153.260564946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-8fwqv" (UID: "6513f255-e88c-4c51-8d5b-17018a5c835d") : secret "networking-console-plugin-cert" not found Apr 22 14:17:19.359292 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.359253 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" event={"ID":"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9","Type":"ContainerStarted","Data":"c0c1bdbea8a22c79e0c75fd7a5b3c4843c4371c9f954b86c2449d49e9f7143e0"} Apr 22 14:17:19.360402 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.360365 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-zkfsf" event={"ID":"727af150-86df-4722-a839-797b4cb1db69","Type":"ContainerStarted","Data":"f0ed4f7b3da55373dc6275ecd6e365066a721d1d1ce79cccbf429ffdd7baed7c"} Apr 22 14:17:19.361899 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.361865 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" event={"ID":"1bdda416-04a6-4a10-a11e-ead137503dd2","Type":"ContainerStarted","Data":"6f2d0bac5f0e3b40b8704576ab9917966982ac9c46b584f36dc7d5bab63451aa"} Apr 22 14:17:19.362022 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.361903 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" event={"ID":"1bdda416-04a6-4a10-a11e-ead137503dd2","Type":"ContainerStarted","Data":"9a39a214a9c8ad601a004e62f19a8729d944420ecf6ac3bef2e8a18e5369880c"} Apr 22 14:17:19.363057 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.363034 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" event={"ID":"36bf42f9-41dd-4d29-9f4c-2472e7df5572","Type":"ContainerStarted","Data":"a4bdc0e74d88b1d79e872e229229b0f8311e8f34b5f41b8d1e8a4e5682e7fa18"} Apr 22 14:17:19.378062 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.378014 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-9m6vn" podStartSLOduration=1.377998481 podStartE2EDuration="1.377998481s" podCreationTimestamp="2026-04-22 14:17:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:17:19.377482613 +0000 UTC m=+152.845360345" watchObservedRunningTime="2026-04-22 14:17:19.377998481 +0000 UTC m=+152.845876208" Apr 22 14:17:19.699293 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.699245 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:19.699484 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:19.699458 2574 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 22 14:17:19.699561 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:19.699522 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls podName:a3973b04-9fff-4f0a-bb01-3de9b19fba5a nodeName:}" failed. No retries permitted until 2026-04-22 14:17:21.699502974 +0000 UTC m=+155.167380683 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-z45rg" (UID: "a3973b04-9fff-4f0a-bb01-3de9b19fba5a") : secret "samples-operator-tls" not found Apr 22 14:17:19.800407 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:19.800366 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:19.800572 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:19.800523 2574 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 22 14:17:19.800626 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:19.800591 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert podName:6513f255-e88c-4c51-8d5b-17018a5c835d nodeName:}" failed. No retries permitted until 2026-04-22 14:17:21.800571219 +0000 UTC m=+155.268448928 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-8fwqv" (UID: "6513f255-e88c-4c51-8d5b-17018a5c835d") : secret "networking-console-plugin-cert" not found Apr 22 14:17:21.716065 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:21.716020 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:21.716533 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:21.716174 2574 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 22 14:17:21.716533 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:21.716248 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls podName:a3973b04-9fff-4f0a-bb01-3de9b19fba5a nodeName:}" failed. No retries permitted until 2026-04-22 14:17:25.716229319 +0000 UTC m=+159.184107043 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-z45rg" (UID: "a3973b04-9fff-4f0a-bb01-3de9b19fba5a") : secret "samples-operator-tls" not found Apr 22 14:17:21.817248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:21.817215 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:21.817353 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:21.817330 2574 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 22 14:17:21.817393 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:21.817386 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert podName:6513f255-e88c-4c51-8d5b-17018a5c835d nodeName:}" failed. No retries permitted until 2026-04-22 14:17:25.81737117 +0000 UTC m=+159.285248878 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-8fwqv" (UID: "6513f255-e88c-4c51-8d5b-17018a5c835d") : secret "networking-console-plugin-cert" not found Apr 22 14:17:22.371065 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.370970 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-zkfsf" event={"ID":"727af150-86df-4722-a839-797b4cb1db69","Type":"ContainerStarted","Data":"e6bc2f7747be75ad033a180b6120f4d20088c3ec0eac917bfa1b9a3c991da0cb"} Apr 22 14:17:22.372293 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.372269 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" event={"ID":"36bf42f9-41dd-4d29-9f4c-2472e7df5572","Type":"ContainerStarted","Data":"40ed85ed8d94878671a089ee4a421e20b969b67d97186696df3bf56a3a1a5e65"} Apr 22 14:17:22.373599 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.373583 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/0.log" Apr 22 14:17:22.373687 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.373615 2574 generic.go:358] "Generic (PLEG): container finished" podID="00643d7f-54c7-4ca9-8d93-e7cf0c7554c9" containerID="c9846b50a1c48b66143e5dd617db7ee9c2824cb48ea29aa7af5fa074cc99f5c7" exitCode=255 Apr 22 14:17:22.373687 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.373642 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" event={"ID":"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9","Type":"ContainerDied","Data":"c9846b50a1c48b66143e5dd617db7ee9c2824cb48ea29aa7af5fa074cc99f5c7"} Apr 22 14:17:22.373920 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.373904 2574 scope.go:117] "RemoveContainer" containerID="c9846b50a1c48b66143e5dd617db7ee9c2824cb48ea29aa7af5fa074cc99f5c7" Apr 22 14:17:22.409926 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.409872 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-operator-585dfdc468-zkfsf" podStartSLOduration=1.200037394 podStartE2EDuration="4.409853118s" podCreationTimestamp="2026-04-22 14:17:18 +0000 UTC" firstStartedPulling="2026-04-22 14:17:18.631061523 +0000 UTC m=+152.098939234" lastFinishedPulling="2026-04-22 14:17:21.840877249 +0000 UTC m=+155.308754958" observedRunningTime="2026-04-22 14:17:22.387937257 +0000 UTC m=+155.855814986" watchObservedRunningTime="2026-04-22 14:17:22.409853118 +0000 UTC m=+155.877730850" Apr 22 14:17:22.527357 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:22.527318 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[registry-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" podUID="d1bd8c4c-012e-4776-934f-9a95a15ce4bb" Apr 22 14:17:22.547492 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:22.547458 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-4kpkf" podUID="0ef3c015-11a0-43bf-b3db-a4ee755ab04b" Apr 22 14:17:22.562643 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:22.562611 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-b6brs" podUID="daaf870a-47fd-4fba-8f87-375c235c1a4c" Apr 22 14:17:22.648025 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.647959 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" podStartSLOduration=1.42883403 podStartE2EDuration="4.647937143s" podCreationTimestamp="2026-04-22 14:17:18 +0000 UTC" firstStartedPulling="2026-04-22 14:17:18.616965692 +0000 UTC m=+152.084843407" lastFinishedPulling="2026-04-22 14:17:21.836068808 +0000 UTC m=+155.303946520" observedRunningTime="2026-04-22 14:17:22.440833295 +0000 UTC m=+155.908711026" watchObservedRunningTime="2026-04-22 14:17:22.647937143 +0000 UTC m=+156.115814874" Apr 22 14:17:22.648903 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.648886 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6"] Apr 22 14:17:22.652311 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.652290 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" Apr 22 14:17:22.660307 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.660261 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-28qkq\"" Apr 22 14:17:22.660307 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.660292 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Apr 22 14:17:22.660460 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.660304 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Apr 22 14:17:22.668597 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.668577 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6"] Apr 22 14:17:22.723307 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.723269 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psn79\" (UniqueName: \"kubernetes.io/projected/c1d8e35d-172e-4b63-9962-341853586a26-kube-api-access-psn79\") pod \"migrator-74bb7799d9-nqqx6\" (UID: \"c1d8e35d-172e-4b63-9962-341853586a26\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" Apr 22 14:17:22.823729 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.823669 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-psn79\" (UniqueName: \"kubernetes.io/projected/c1d8e35d-172e-4b63-9962-341853586a26-kube-api-access-psn79\") pod \"migrator-74bb7799d9-nqqx6\" (UID: \"c1d8e35d-172e-4b63-9962-341853586a26\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" Apr 22 14:17:22.831661 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.831632 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-psn79\" (UniqueName: \"kubernetes.io/projected/c1d8e35d-172e-4b63-9962-341853586a26-kube-api-access-psn79\") pod \"migrator-74bb7799d9-nqqx6\" (UID: \"c1d8e35d-172e-4b63-9962-341853586a26\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" Apr 22 14:17:22.961122 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:22.961031 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" Apr 22 14:17:23.082397 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.082367 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6"] Apr 22 14:17:23.086011 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:23.085982 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1d8e35d_172e_4b63_9962_341853586a26.slice/crio-3054c623ef9538956caf77ce45ab9e3f3be6c1c5974c79a3d973761cae4718be WatchSource:0}: Error finding container 3054c623ef9538956caf77ce45ab9e3f3be6c1c5974c79a3d973761cae4718be: Status 404 returned error can't find the container with id 3054c623ef9538956caf77ce45ab9e3f3be6c1c5974c79a3d973761cae4718be Apr 22 14:17:23.378880 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.378035 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:17:23.379035 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.379000 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/0.log" Apr 22 14:17:23.379089 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.379033 2574 generic.go:358] "Generic (PLEG): container finished" podID="00643d7f-54c7-4ca9-8d93-e7cf0c7554c9" containerID="b46eec37b453961d1728e6c1f7b59c173e4c85ab4d90d300a8d60a2e212a7b1e" exitCode=255 Apr 22 14:17:23.379143 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.379122 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" event={"ID":"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9","Type":"ContainerDied","Data":"b46eec37b453961d1728e6c1f7b59c173e4c85ab4d90d300a8d60a2e212a7b1e"} Apr 22 14:17:23.379198 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.379163 2574 scope.go:117] "RemoveContainer" containerID="c9846b50a1c48b66143e5dd617db7ee9c2824cb48ea29aa7af5fa074cc99f5c7" Apr 22 14:17:23.379611 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.379579 2574 scope.go:117] "RemoveContainer" containerID="b46eec37b453961d1728e6c1f7b59c173e4c85ab4d90d300a8d60a2e212a7b1e" Apr 22 14:17:23.379865 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:23.379835 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-pxmjh_openshift-console-operator(00643d7f-54c7-4ca9-8d93-e7cf0c7554c9)\"" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" podUID="00643d7f-54c7-4ca9-8d93-e7cf0c7554c9" Apr 22 14:17:23.381268 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.381251 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:17:23.381474 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.381352 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" event={"ID":"c1d8e35d-172e-4b63-9962-341853586a26","Type":"ContainerStarted","Data":"3054c623ef9538956caf77ce45ab9e3f3be6c1c5974c79a3d973761cae4718be"} Apr 22 14:17:23.381474 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:23.381465 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4kpkf" Apr 22 14:17:24.003800 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:24.003759 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-certs], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-multus/network-metrics-daemon-p65qg" podUID="1eca98e2-015f-4ce7-bcd6-76878fa26d4b" Apr 22 14:17:24.051313 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.051291 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-865cb79987-pvglc"] Apr 22 14:17:24.054347 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.054192 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.056598 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.056577 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Apr 22 14:17:24.056723 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.056645 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Apr 22 14:17:24.056803 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.056779 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Apr 22 14:17:24.056861 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.056848 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Apr 22 14:17:24.057485 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.057467 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-npsnt\"" Apr 22 14:17:24.062105 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.062084 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-pvglc"] Apr 22 14:17:24.135556 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.135534 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f735dec2-affd-481d-9c3d-295dbab72362-signing-cabundle\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.135660 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.135603 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f735dec2-affd-481d-9c3d-295dbab72362-signing-key\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.135739 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.135697 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkhhh\" (UniqueName: \"kubernetes.io/projected/f735dec2-affd-481d-9c3d-295dbab72362-kube-api-access-tkhhh\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.236193 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.236160 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f735dec2-affd-481d-9c3d-295dbab72362-signing-cabundle\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.236381 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.236229 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f735dec2-affd-481d-9c3d-295dbab72362-signing-key\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.236381 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.236249 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tkhhh\" (UniqueName: \"kubernetes.io/projected/f735dec2-affd-481d-9c3d-295dbab72362-kube-api-access-tkhhh\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.236988 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.236939 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f735dec2-affd-481d-9c3d-295dbab72362-signing-cabundle\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.238789 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.238772 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f735dec2-affd-481d-9c3d-295dbab72362-signing-key\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.244575 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.244554 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkhhh\" (UniqueName: \"kubernetes.io/projected/f735dec2-affd-481d-9c3d-295dbab72362-kube-api-access-tkhhh\") pod \"service-ca-865cb79987-pvglc\" (UID: \"f735dec2-affd-481d-9c3d-295dbab72362\") " pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.384216 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.384185 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-pvglc" Apr 22 14:17:24.385414 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.385380 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" event={"ID":"c1d8e35d-172e-4b63-9962-341853586a26","Type":"ContainerStarted","Data":"37e336583cbbd5c2bcbbd964758395bba6bc5ccc380fbaca1af81767a946571b"} Apr 22 14:17:24.385414 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.385418 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" event={"ID":"c1d8e35d-172e-4b63-9962-341853586a26","Type":"ContainerStarted","Data":"333c3f684d14cc684178fd65cd10159fa14579738e8d686432ae89892fcf5b91"} Apr 22 14:17:24.386714 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.386692 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:17:24.387005 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.386993 2574 scope.go:117] "RemoveContainer" containerID="b46eec37b453961d1728e6c1f7b59c173e4c85ab4d90d300a8d60a2e212a7b1e" Apr 22 14:17:24.387150 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:24.387133 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-pxmjh_openshift-console-operator(00643d7f-54c7-4ca9-8d93-e7cf0c7554c9)\"" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" podUID="00643d7f-54c7-4ca9-8d93-e7cf0c7554c9" Apr 22 14:17:24.405766 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.404465 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-nqqx6" podStartSLOduration=1.449451221 podStartE2EDuration="2.404447037s" podCreationTimestamp="2026-04-22 14:17:22 +0000 UTC" firstStartedPulling="2026-04-22 14:17:23.088336659 +0000 UTC m=+156.556214374" lastFinishedPulling="2026-04-22 14:17:24.04333248 +0000 UTC m=+157.511210190" observedRunningTime="2026-04-22 14:17:24.402135103 +0000 UTC m=+157.870012834" watchObservedRunningTime="2026-04-22 14:17:24.404447037 +0000 UTC m=+157.872324770" Apr 22 14:17:24.509411 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:24.509388 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-pvglc"] Apr 22 14:17:24.510995 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:24.510964 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf735dec2_affd_481d_9c3d_295dbab72362.slice/crio-58ab901c920aed7a860a408b14e1f87c5803ef8c8326495f2dbcde8e2c35f949 WatchSource:0}: Error finding container 58ab901c920aed7a860a408b14e1f87c5803ef8c8326495f2dbcde8e2c35f949: Status 404 returned error can't find the container with id 58ab901c920aed7a860a408b14e1f87c5803ef8c8326495f2dbcde8e2c35f949 Apr 22 14:17:25.391162 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:25.391127 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-pvglc" event={"ID":"f735dec2-affd-481d-9c3d-295dbab72362","Type":"ContainerStarted","Data":"58ab901c920aed7a860a408b14e1f87c5803ef8c8326495f2dbcde8e2c35f949"} Apr 22 14:17:25.749140 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:25.749058 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:25.749303 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:25.749210 2574 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 22 14:17:25.749303 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:25.749281 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls podName:a3973b04-9fff-4f0a-bb01-3de9b19fba5a nodeName:}" failed. No retries permitted until 2026-04-22 14:17:33.749261607 +0000 UTC m=+167.217139319 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-z45rg" (UID: "a3973b04-9fff-4f0a-bb01-3de9b19fba5a") : secret "samples-operator-tls" not found Apr 22 14:17:25.849821 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:25.849780 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:25.849992 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:25.849968 2574 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 22 14:17:25.850064 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:25.850050 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert podName:6513f255-e88c-4c51-8d5b-17018a5c835d nodeName:}" failed. No retries permitted until 2026-04-22 14:17:33.850029949 +0000 UTC m=+167.317907664 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-8fwqv" (UID: "6513f255-e88c-4c51-8d5b-17018a5c835d") : secret "networking-console-plugin-cert" not found Apr 22 14:17:26.014498 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:26.014477 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-d5pcf_cd5f393a-8e7d-4fec-ba9c-362d5579a518/dns-node-resolver/0.log" Apr 22 14:17:26.395324 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:26.395291 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-pvglc" event={"ID":"f735dec2-affd-481d-9c3d-295dbab72362","Type":"ContainerStarted","Data":"c359a873d719be89859ea9e2da6066c2d533c639b24a1e0b08308fbac179cc01"} Apr 22 14:17:26.413464 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:26.413418 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-865cb79987-pvglc" podStartSLOduration=0.914438722 podStartE2EDuration="2.413401872s" podCreationTimestamp="2026-04-22 14:17:24 +0000 UTC" firstStartedPulling="2026-04-22 14:17:24.512733052 +0000 UTC m=+157.980610760" lastFinishedPulling="2026-04-22 14:17:26.011696184 +0000 UTC m=+159.479573910" observedRunningTime="2026-04-22 14:17:26.412556755 +0000 UTC m=+159.880434485" watchObservedRunningTime="2026-04-22 14:17:26.413401872 +0000 UTC m=+159.881279600" Apr 22 14:17:27.216213 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:27.216183 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-dx5kd_3688a996-ab5a-4d60-b8e0-19b8a75bae48/node-ca/0.log" Apr 22 14:17:27.465085 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:27.465040 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:17:27.465512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:27.465131 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") pod \"image-registry-7bfcd86c87-jl5cs\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:17:27.465512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:27.465189 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:17:27.465512 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:27.465336 2574 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 14:17:27.465512 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:27.465393 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls podName:0ef3c015-11a0-43bf-b3db-a4ee755ab04b nodeName:}" failed. No retries permitted until 2026-04-22 14:19:29.465375095 +0000 UTC m=+282.933252808 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls") pod "dns-default-4kpkf" (UID: "0ef3c015-11a0-43bf-b3db-a4ee755ab04b") : secret "dns-default-metrics-tls" not found Apr 22 14:17:27.466362 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:27.466297 2574 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 14:17:27.466929 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:27.466917 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert podName:daaf870a-47fd-4fba-8f87-375c235c1a4c nodeName:}" failed. No retries permitted until 2026-04-22 14:19:29.466901678 +0000 UTC m=+282.934779393 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert") pod "ingress-canary-b6brs" (UID: "daaf870a-47fd-4fba-8f87-375c235c1a4c") : secret "canary-serving-cert" not found Apr 22 14:17:27.467094 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:27.466827 2574 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 22 14:17:27.467175 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:27.467167 2574 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-7bfcd86c87-jl5cs: secret "image-registry-tls" not found Apr 22 14:17:27.467273 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:27.467265 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls podName:d1bd8c4c-012e-4776-934f-9a95a15ce4bb nodeName:}" failed. No retries permitted until 2026-04-22 14:19:29.46725178 +0000 UTC m=+282.935129501 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls") pod "image-registry-7bfcd86c87-jl5cs" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb") : secret "image-registry-tls" not found Apr 22 14:17:28.409734 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:28.409695 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-nqqx6_c1d8e35d-172e-4b63-9962-341853586a26/migrator/0.log" Apr 22 14:17:28.607419 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:28.607367 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:28.607419 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:28.607424 2574 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:28.607857 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:28.607835 2574 scope.go:117] "RemoveContainer" containerID="b46eec37b453961d1728e6c1f7b59c173e4c85ab4d90d300a8d60a2e212a7b1e" Apr 22 14:17:28.608015 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:28.607998 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-pxmjh_openshift-console-operator(00643d7f-54c7-4ca9-8d93-e7cf0c7554c9)\"" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" podUID="00643d7f-54c7-4ca9-8d93-e7cf0c7554c9" Apr 22 14:17:28.608658 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:28.608641 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-nqqx6_c1d8e35d-172e-4b63-9962-341853586a26/graceful-termination/0.log" Apr 22 14:17:28.814750 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:28.814657 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-9lcw9_36bf42f9-41dd-4d29-9f4c-2472e7df5572/kube-storage-version-migrator-operator/0.log" Apr 22 14:17:33.825476 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:33.825393 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:33.827954 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:33.827935 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3973b04-9fff-4f0a-bb01-3de9b19fba5a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-z45rg\" (UID: \"a3973b04-9fff-4f0a-bb01-3de9b19fba5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:33.925898 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:33.925859 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:33.926054 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:33.926013 2574 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 22 14:17:33.926100 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:17:33.926078 2574 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert podName:6513f255-e88c-4c51-8d5b-17018a5c835d nodeName:}" failed. No retries permitted until 2026-04-22 14:17:49.926061514 +0000 UTC m=+183.393939227 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-8fwqv" (UID: "6513f255-e88c-4c51-8d5b-17018a5c835d") : secret "networking-console-plugin-cert" not found Apr 22 14:17:33.972715 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:33.972659 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" Apr 22 14:17:34.090915 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:34.090827 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg"] Apr 22 14:17:34.416753 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:34.416715 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" event={"ID":"a3973b04-9fff-4f0a-bb01-3de9b19fba5a","Type":"ContainerStarted","Data":"b43076435c65ae9b8c5a54c04124762df24398d47f848a7c6c65fc453b52fd01"} Apr 22 14:17:34.984031 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:34.983996 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:17:36.423410 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:36.423374 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" event={"ID":"a3973b04-9fff-4f0a-bb01-3de9b19fba5a","Type":"ContainerStarted","Data":"22197758e774cdee5ec46c315bc708bae4aca45840b2a6f26387c2136749c254"} Apr 22 14:17:36.423410 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:36.423409 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" event={"ID":"a3973b04-9fff-4f0a-bb01-3de9b19fba5a","Type":"ContainerStarted","Data":"adc52641f5500d42487d6a0e6133ea9dd055695ae6be67e6787663d14bc63bdb"} Apr 22 14:17:36.439920 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:36.439862 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-z45rg" podStartSLOduration=16.912398451 podStartE2EDuration="18.439842983s" podCreationTimestamp="2026-04-22 14:17:18 +0000 UTC" firstStartedPulling="2026-04-22 14:17:34.155570826 +0000 UTC m=+167.623448534" lastFinishedPulling="2026-04-22 14:17:35.683015343 +0000 UTC m=+169.150893066" observedRunningTime="2026-04-22 14:17:36.43859509 +0000 UTC m=+169.906472820" watchObservedRunningTime="2026-04-22 14:17:36.439842983 +0000 UTC m=+169.907720713" Apr 22 14:17:38.983847 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:38.983799 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:17:38.984272 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:38.984206 2574 scope.go:117] "RemoveContainer" containerID="b46eec37b453961d1728e6c1f7b59c173e4c85ab4d90d300a8d60a2e212a7b1e" Apr 22 14:17:39.432973 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:39.432947 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:17:39.433130 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:39.433019 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" event={"ID":"00643d7f-54c7-4ca9-8d93-e7cf0c7554c9","Type":"ContainerStarted","Data":"00f7abe61b2690ccf04a48a3daa849321c543d7a0a762400682357e3a2331acb"} Apr 22 14:17:39.433311 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:39.433282 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:39.450459 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:39.450415 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" podStartSLOduration=18.345462888 podStartE2EDuration="21.450401652s" podCreationTimestamp="2026-04-22 14:17:18 +0000 UTC" firstStartedPulling="2026-04-22 14:17:18.733375376 +0000 UTC m=+152.201253084" lastFinishedPulling="2026-04-22 14:17:21.838314124 +0000 UTC m=+155.306191848" observedRunningTime="2026-04-22 14:17:39.449244906 +0000 UTC m=+172.917122640" watchObservedRunningTime="2026-04-22 14:17:39.450401652 +0000 UTC m=+172.918279383" Apr 22 14:17:39.480452 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:39.480422 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-9d4b6777b-pxmjh" Apr 22 14:17:44.740546 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.740508 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-588c45675c-6nbl6"] Apr 22 14:17:44.743765 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.743732 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-2k8nd"] Apr 22 14:17:44.743949 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.743930 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.747326 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.747305 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.747876 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.747859 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 22 14:17:44.748099 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.748084 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 22 14:17:44.748179 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.748161 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 22 14:17:44.748238 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.748161 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-78jk5\"" Apr 22 14:17:44.750158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.750138 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 22 14:17:44.759955 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.759927 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 22 14:17:44.760067 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.760026 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 22 14:17:44.760067 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.760026 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 22 14:17:44.760177 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.760051 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 22 14:17:44.760250 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.760236 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 22 14:17:44.760331 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.760315 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-4zp5l\"" Apr 22 14:17:44.790323 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.790292 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-2k8nd"] Apr 22 14:17:44.823767 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823730 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-config\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.823942 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823784 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-service-ca\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.823942 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823807 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x8dr\" (UniqueName: \"kubernetes.io/projected/3dffc553-a2b4-4156-af72-d9a25248a386-kube-api-access-2x8dr\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.823942 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823824 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-serving-cert\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.823942 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823844 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-oauth-config\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.823942 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823923 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/3dffc553-a2b4-4156-af72-d9a25248a386-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.824158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823960 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/3dffc553-a2b4-4156-af72-d9a25248a386-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.824158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.823988 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtsfs\" (UniqueName: \"kubernetes.io/projected/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-kube-api-access-wtsfs\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.824158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.824048 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/3dffc553-a2b4-4156-af72-d9a25248a386-crio-socket\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.824158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.824088 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-oauth-serving-cert\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.824158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.824121 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/3dffc553-a2b4-4156-af72-d9a25248a386-data-volume\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.885356 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.885323 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-588c45675c-6nbl6"] Apr 22 14:17:44.924917 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.924885 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/3dffc553-a2b4-4156-af72-d9a25248a386-data-volume\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925075 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.924964 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-config\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.925075 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.924993 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-service-ca\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.925075 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925018 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2x8dr\" (UniqueName: \"kubernetes.io/projected/3dffc553-a2b4-4156-af72-d9a25248a386-kube-api-access-2x8dr\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925075 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925048 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-serving-cert\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.925260 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925078 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-oauth-config\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.925260 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925105 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/3dffc553-a2b4-4156-af72-d9a25248a386-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925329 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925252 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/3dffc553-a2b4-4156-af72-d9a25248a386-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925329 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925293 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/3dffc553-a2b4-4156-af72-d9a25248a386-data-volume\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925329 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925301 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wtsfs\" (UniqueName: \"kubernetes.io/projected/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-kube-api-access-wtsfs\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.925419 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925343 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/3dffc553-a2b4-4156-af72-d9a25248a386-crio-socket\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925419 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925388 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-oauth-serving-cert\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.925586 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925566 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/3dffc553-a2b4-4156-af72-d9a25248a386-crio-socket\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925643 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925629 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/3dffc553-a2b4-4156-af72-d9a25248a386-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.925800 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925783 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-config\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.925856 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.925814 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-service-ca\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.926165 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.926108 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-oauth-serving-cert\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.927758 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.927738 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/3dffc553-a2b4-4156-af72-d9a25248a386-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:44.927872 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.927849 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-oauth-config\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.928062 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.928044 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-serving-cert\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.942930 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.942902 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtsfs\" (UniqueName: \"kubernetes.io/projected/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-kube-api-access-wtsfs\") pod \"console-588c45675c-6nbl6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:44.949602 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:44.949578 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x8dr\" (UniqueName: \"kubernetes.io/projected/3dffc553-a2b4-4156-af72-d9a25248a386-kube-api-access-2x8dr\") pod \"insights-runtime-extractor-2k8nd\" (UID: \"3dffc553-a2b4-4156-af72-d9a25248a386\") " pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:45.054547 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:45.054460 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:45.058292 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:45.058268 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-2k8nd" Apr 22 14:17:45.224108 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:45.224077 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-2k8nd"] Apr 22 14:17:45.224276 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:45.224250 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dffc553_a2b4_4156_af72_d9a25248a386.slice/crio-2ae822778e4a307c6ecbfae1c529224572fc2512ab42f512706fcddec747b236 WatchSource:0}: Error finding container 2ae822778e4a307c6ecbfae1c529224572fc2512ab42f512706fcddec747b236: Status 404 returned error can't find the container with id 2ae822778e4a307c6ecbfae1c529224572fc2512ab42f512706fcddec747b236 Apr 22 14:17:45.224825 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:45.224805 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb79bdabc_7a55_43ee_a856_0cbc5b1c49c6.slice/crio-7f295c412dedc31c663ed330c2e8f8e0f80315ea36a5c2f9fa7dce39568520d8 WatchSource:0}: Error finding container 7f295c412dedc31c663ed330c2e8f8e0f80315ea36a5c2f9fa7dce39568520d8: Status 404 returned error can't find the container with id 7f295c412dedc31c663ed330c2e8f8e0f80315ea36a5c2f9fa7dce39568520d8 Apr 22 14:17:45.228358 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:45.228337 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-588c45675c-6nbl6"] Apr 22 14:17:45.448129 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:45.448095 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-2k8nd" event={"ID":"3dffc553-a2b4-4156-af72-d9a25248a386","Type":"ContainerStarted","Data":"24d8fd6b880992674103ae29e2a192f557437c35d9a129264b1fa73105b1b850"} Apr 22 14:17:45.448129 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:45.448132 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-2k8nd" event={"ID":"3dffc553-a2b4-4156-af72-d9a25248a386","Type":"ContainerStarted","Data":"2ae822778e4a307c6ecbfae1c529224572fc2512ab42f512706fcddec747b236"} Apr 22 14:17:45.449079 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:45.449050 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-588c45675c-6nbl6" event={"ID":"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6","Type":"ContainerStarted","Data":"7f295c412dedc31c663ed330c2e8f8e0f80315ea36a5c2f9fa7dce39568520d8"} Apr 22 14:17:46.455051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:46.454967 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-2k8nd" event={"ID":"3dffc553-a2b4-4156-af72-d9a25248a386","Type":"ContainerStarted","Data":"8d07fe4da08d7e7cc7e94ae8d8fb7e9d30acba05a5840c545933450952e95b0f"} Apr 22 14:17:48.461567 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:48.461476 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-588c45675c-6nbl6" event={"ID":"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6","Type":"ContainerStarted","Data":"6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e"} Apr 22 14:17:48.463390 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:48.463367 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-2k8nd" event={"ID":"3dffc553-a2b4-4156-af72-d9a25248a386","Type":"ContainerStarted","Data":"4fcd1ba7317258be266ef4b148c5450fc748811f49755ac4863d98298f58b5f3"} Apr 22 14:17:48.481620 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:48.481568 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-588c45675c-6nbl6" podStartSLOduration=1.494812921 podStartE2EDuration="4.481554886s" podCreationTimestamp="2026-04-22 14:17:44 +0000 UTC" firstStartedPulling="2026-04-22 14:17:45.22664081 +0000 UTC m=+178.694518519" lastFinishedPulling="2026-04-22 14:17:48.213382763 +0000 UTC m=+181.681260484" observedRunningTime="2026-04-22 14:17:48.480720209 +0000 UTC m=+181.948597939" watchObservedRunningTime="2026-04-22 14:17:48.481554886 +0000 UTC m=+181.949432614" Apr 22 14:17:48.501173 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:48.501125 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-2k8nd" podStartSLOduration=1.576481601 podStartE2EDuration="4.501109326s" podCreationTimestamp="2026-04-22 14:17:44 +0000 UTC" firstStartedPulling="2026-04-22 14:17:45.288687806 +0000 UTC m=+178.756565528" lastFinishedPulling="2026-04-22 14:17:48.213315529 +0000 UTC m=+181.681193253" observedRunningTime="2026-04-22 14:17:48.500545321 +0000 UTC m=+181.968423053" watchObservedRunningTime="2026-04-22 14:17:48.501109326 +0000 UTC m=+181.968987055" Apr 22 14:17:49.973358 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:49.973316 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:49.976202 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:49.976175 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6513f255-e88c-4c51-8d5b-17018a5c835d-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-8fwqv\" (UID: \"6513f255-e88c-4c51-8d5b-17018a5c835d\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:49.993276 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:49.993245 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"default-dockercfg-xnxkr\"" Apr 22 14:17:50.001277 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:50.001258 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" Apr 22 14:17:50.120465 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:50.120432 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv"] Apr 22 14:17:50.124029 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:50.124004 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6513f255_e88c_4c51_8d5b_17018a5c835d.slice/crio-766d01bc848f8efcfcb878906c47dfff864098f3a0ac5d097918124d9203c16e WatchSource:0}: Error finding container 766d01bc848f8efcfcb878906c47dfff864098f3a0ac5d097918124d9203c16e: Status 404 returned error can't find the container with id 766d01bc848f8efcfcb878906c47dfff864098f3a0ac5d097918124d9203c16e Apr 22 14:17:50.469345 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:50.469304 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" event={"ID":"6513f255-e88c-4c51-8d5b-17018a5c835d","Type":"ContainerStarted","Data":"766d01bc848f8efcfcb878906c47dfff864098f3a0ac5d097918124d9203c16e"} Apr 22 14:17:51.473005 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:51.472965 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" event={"ID":"6513f255-e88c-4c51-8d5b-17018a5c835d","Type":"ContainerStarted","Data":"8ad35d19d9a4de03a15fd1890c8619787234cadbb12ff78f4eb434efa29de3ed"} Apr 22 14:17:51.488026 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:51.487955 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-console/networking-console-plugin-cb95c66f6-8fwqv" podStartSLOduration=32.512189481 podStartE2EDuration="33.487940969s" podCreationTimestamp="2026-04-22 14:17:18 +0000 UTC" firstStartedPulling="2026-04-22 14:17:50.126382634 +0000 UTC m=+183.594260343" lastFinishedPulling="2026-04-22 14:17:51.102134111 +0000 UTC m=+184.570011831" observedRunningTime="2026-04-22 14:17:51.487298957 +0000 UTC m=+184.955176698" watchObservedRunningTime="2026-04-22 14:17:51.487940969 +0000 UTC m=+184.955818698" Apr 22 14:17:52.013297 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.013264 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-7df65c58c4-27ft2"] Apr 22 14:17:52.016614 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.016596 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.024469 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.024446 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 22 14:17:52.032608 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.032580 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7df65c58c4-27ft2"] Apr 22 14:17:52.194024 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.193983 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-oauth-serving-cert\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.194024 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.194030 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhvtd\" (UniqueName: \"kubernetes.io/projected/c6863efe-089d-4c10-a59b-7cc11a649cc1-kube-api-access-fhvtd\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.194237 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.194091 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-config\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.194237 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.194125 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-service-ca\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.194237 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.194144 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-oauth-config\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.194237 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.194164 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-trusted-ca-bundle\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.194237 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.194226 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-serving-cert\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295174 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295098 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-oauth-serving-cert\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295174 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295137 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fhvtd\" (UniqueName: \"kubernetes.io/projected/c6863efe-089d-4c10-a59b-7cc11a649cc1-kube-api-access-fhvtd\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295174 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295160 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-config\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295410 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295180 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-service-ca\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295410 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295208 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-oauth-config\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295410 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295235 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-trusted-ca-bundle\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295410 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295276 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-serving-cert\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295936 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295903 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-oauth-serving-cert\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.295936 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295923 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-config\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.296273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.295949 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-service-ca\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.296273 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.296139 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-trusted-ca-bundle\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.297813 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.297777 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-serving-cert\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.297918 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.297870 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-oauth-config\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.304289 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.304269 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhvtd\" (UniqueName: \"kubernetes.io/projected/c6863efe-089d-4c10-a59b-7cc11a649cc1-kube-api-access-fhvtd\") pod \"console-7df65c58c4-27ft2\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.326352 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.326330 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:17:52.452185 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.452154 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7df65c58c4-27ft2"] Apr 22 14:17:52.455375 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:17:52.455345 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6863efe_089d_4c10_a59b_7cc11a649cc1.slice/crio-7310873f1ae2711bed8313cb923742aaf959c1034a5c21cc27e67ae7855e5d69 WatchSource:0}: Error finding container 7310873f1ae2711bed8313cb923742aaf959c1034a5c21cc27e67ae7855e5d69: Status 404 returned error can't find the container with id 7310873f1ae2711bed8313cb923742aaf959c1034a5c21cc27e67ae7855e5d69 Apr 22 14:17:52.476255 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:52.476228 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7df65c58c4-27ft2" event={"ID":"c6863efe-089d-4c10-a59b-7cc11a649cc1","Type":"ContainerStarted","Data":"7310873f1ae2711bed8313cb923742aaf959c1034a5c21cc27e67ae7855e5d69"} Apr 22 14:17:53.480773 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:53.480737 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7df65c58c4-27ft2" event={"ID":"c6863efe-089d-4c10-a59b-7cc11a649cc1","Type":"ContainerStarted","Data":"b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e"} Apr 22 14:17:53.499341 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:53.499279 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7df65c58c4-27ft2" podStartSLOduration=2.499262819 podStartE2EDuration="2.499262819s" podCreationTimestamp="2026-04-22 14:17:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:17:53.499079351 +0000 UTC m=+186.966957072" watchObservedRunningTime="2026-04-22 14:17:53.499262819 +0000 UTC m=+186.967140550" Apr 22 14:17:55.054985 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:55.054949 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:55.054985 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:55.054995 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:55.059569 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:55.059545 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:17:55.490725 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:17:55.490698 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:18:02.327116 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:02.327080 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:18:02.327501 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:02.327124 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:18:02.331978 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:02.331954 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:18:02.512836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:02.512807 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:18:02.586792 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:02.586701 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-588c45675c-6nbl6"] Apr 22 14:18:06.496363 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.496331 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-7bfcd86c87-jl5cs"] Apr 22 14:18:06.496785 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:18:06.496552 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[registry-tls], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" podUID="d1bd8c4c-012e-4776-934f-9a95a15ce4bb" Apr 22 14:18:06.523238 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.523209 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:18:06.528032 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.528012 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:18:06.628490 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628457 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-ca-trust-extracted\") pod \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " Apr 22 14:18:06.628697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628502 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-certificates\") pod \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " Apr 22 14:18:06.628697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628530 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-bound-sa-token\") pod \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " Apr 22 14:18:06.628697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628589 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-installation-pull-secrets\") pod \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " Apr 22 14:18:06.628697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628605 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-image-registry-private-configuration\") pod \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " Apr 22 14:18:06.628697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628624 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-trusted-ca\") pod \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " Apr 22 14:18:06.628697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628659 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khslj\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-kube-api-access-khslj\") pod \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\" (UID: \"d1bd8c4c-012e-4776-934f-9a95a15ce4bb\") " Apr 22 14:18:06.628991 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628766 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d1bd8c4c-012e-4776-934f-9a95a15ce4bb" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 14:18:06.629049 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.628987 2574 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-ca-trust-extracted\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:06.629099 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.629043 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d1bd8c4c-012e-4776-934f-9a95a15ce4bb" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:06.629099 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.629081 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d1bd8c4c-012e-4776-934f-9a95a15ce4bb" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:06.631717 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.631665 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-kube-api-access-khslj" (OuterVolumeSpecName: "kube-api-access-khslj") pod "d1bd8c4c-012e-4776-934f-9a95a15ce4bb" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb"). InnerVolumeSpecName "kube-api-access-khslj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 14:18:06.631717 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.631662 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "d1bd8c4c-012e-4776-934f-9a95a15ce4bb" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 14:18:06.631889 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.631735 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d1bd8c4c-012e-4776-934f-9a95a15ce4bb" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 14:18:06.631889 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.631792 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d1bd8c4c-012e-4776-934f-9a95a15ce4bb" (UID: "d1bd8c4c-012e-4776-934f-9a95a15ce4bb"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 14:18:06.730344 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.730308 2574 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-trusted-ca\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:06.730344 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.730340 2574 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-khslj\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-kube-api-access-khslj\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:06.730344 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.730352 2574 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-certificates\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:06.730559 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.730361 2574 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-bound-sa-token\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:06.730559 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.730371 2574 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-installation-pull-secrets\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:06.730559 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:06.730380 2574 reconciler_common.go:299] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-image-registry-private-configuration\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:07.525592 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:07.525558 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-7bfcd86c87-jl5cs" Apr 22 14:18:07.570749 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:07.570714 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-7bfcd86c87-jl5cs"] Apr 22 14:18:07.577645 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:07.577620 2574 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-7bfcd86c87-jl5cs"] Apr 22 14:18:07.637724 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:07.637691 2574 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d1bd8c4c-012e-4776-934f-9a95a15ce4bb-registry-tls\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:08.592242 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.592207 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8"] Apr 22 14:18:08.595508 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.595487 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.598226 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.598203 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-kube-rbac-proxy-config\"" Apr 22 14:18:08.599226 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.599206 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 22 14:18:08.599335 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.599213 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-dockercfg-ngntk\"" Apr 22 14:18:08.603856 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.603837 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 22 14:18:08.604209 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.604196 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-tls\"" Apr 22 14:18:08.604450 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.604433 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 22 14:18:08.613566 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.613546 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8"] Apr 22 14:18:08.646760 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.646721 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/82e2b9ba-996b-4f92-b967-0052faa65d73-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.646910 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.646784 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czgt5\" (UniqueName: \"kubernetes.io/projected/82e2b9ba-996b-4f92-b967-0052faa65d73-kube-api-access-czgt5\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.646910 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.646861 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/82e2b9ba-996b-4f92-b967-0052faa65d73-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.646993 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.646973 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/82e2b9ba-996b-4f92-b967-0052faa65d73-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.714212 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.714175 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-69db897b98-sjd66"] Apr 22 14:18:08.716352 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.716329 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.718743 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.718721 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-kube-rbac-proxy-config\"" Apr 22 14:18:08.718866 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.718771 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-dockercfg-w84t6\"" Apr 22 14:18:08.719106 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.719090 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-custom-resource-state-configmap\"" Apr 22 14:18:08.720565 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.720546 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-tls\"" Apr 22 14:18:08.729109 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.729091 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-bzjx9"] Apr 22 14:18:08.731309 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.731278 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.731696 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.731634 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-69db897b98-sjd66"] Apr 22 14:18:08.733487 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.733465 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 22 14:18:08.733628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.733495 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 22 14:18:08.733628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.733466 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 22 14:18:08.733628 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.733477 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-2pzx2\"" Apr 22 14:18:08.747840 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.747814 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/eeeda31a-52ab-4389-b06c-eea0f137d09f-volume-directive-shadow\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.747916 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.747842 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-tls\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.747916 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.747858 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eeeda31a-52ab-4389-b06c-eea0f137d09f-metrics-client-ca\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.747990 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.747911 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.747990 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.747949 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/82e2b9ba-996b-4f92-b967-0052faa65d73-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.747990 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.747974 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nlxp\" (UniqueName: \"kubernetes.io/projected/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-api-access-9nlxp\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.748099 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.748086 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/82e2b9ba-996b-4f92-b967-0052faa65d73-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.748135 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.748116 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/82e2b9ba-996b-4f92-b967-0052faa65d73-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.748167 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.748136 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.748167 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.748160 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-czgt5\" (UniqueName: \"kubernetes.io/projected/82e2b9ba-996b-4f92-b967-0052faa65d73-kube-api-access-czgt5\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.748823 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.748795 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/82e2b9ba-996b-4f92-b967-0052faa65d73-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.750491 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.750468 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/82e2b9ba-996b-4f92-b967-0052faa65d73-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.750491 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.750479 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/82e2b9ba-996b-4f92-b967-0052faa65d73-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.761992 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.761969 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-czgt5\" (UniqueName: \"kubernetes.io/projected/82e2b9ba-996b-4f92-b967-0052faa65d73-kube-api-access-czgt5\") pod \"openshift-state-metrics-9d44df66c-lvzv8\" (UID: \"82e2b9ba-996b-4f92-b967-0052faa65d73\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.849424 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849334 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.849424 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849392 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-accelerators-collector-config\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.849604 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849519 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-wtmp\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.849604 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849564 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d148eabc-a2af-41d8-ba7d-039b9da0132a-metrics-client-ca\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.849664 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849603 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/eeeda31a-52ab-4389-b06c-eea0f137d09f-volume-directive-shadow\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.849664 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849648 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-tls\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.849788 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849707 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eeeda31a-52ab-4389-b06c-eea0f137d09f-metrics-client-ca\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.849788 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849729 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqgvc\" (UniqueName: \"kubernetes.io/projected/d148eabc-a2af-41d8-ba7d-039b9da0132a-kube-api-access-pqgvc\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.849891 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849872 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.849935 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849907 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-textfile\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.849984 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.849958 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/eeeda31a-52ab-4389-b06c-eea0f137d09f-volume-directive-shadow\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.850037 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.850008 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9nlxp\" (UniqueName: \"kubernetes.io/projected/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-api-access-9nlxp\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.850084 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.850048 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-sys\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.850140 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.850107 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-root\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.850188 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.850133 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.850188 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.850167 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.850188 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.850183 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-tls\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.850377 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.850353 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eeeda31a-52ab-4389-b06c-eea0f137d09f-metrics-client-ca\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.852386 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.852357 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-tls\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.852488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.852464 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.858909 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.858887 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nlxp\" (UniqueName: \"kubernetes.io/projected/eeeda31a-52ab-4389-b06c-eea0f137d09f-kube-api-access-9nlxp\") pod \"kube-state-metrics-69db897b98-sjd66\" (UID: \"eeeda31a-52ab-4389-b06c-eea0f137d09f\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:08.906538 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.906504 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" Apr 22 14:18:08.951311 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951280 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pqgvc\" (UniqueName: \"kubernetes.io/projected/d148eabc-a2af-41d8-ba7d-039b9da0132a-kube-api-access-pqgvc\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951483 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951372 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-textfile\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951483 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951438 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-sys\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951594 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951487 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-root\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951594 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951510 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951594 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951551 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-tls\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951592 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-accelerators-collector-config\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951644 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-wtmp\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951699 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-root\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951776 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951713 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d148eabc-a2af-41d8-ba7d-039b9da0132a-metrics-client-ca\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.951969 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.951878 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-sys\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.952051 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.952023 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-textfile\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.952376 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.952339 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-wtmp\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.952936 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.952911 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d148eabc-a2af-41d8-ba7d-039b9da0132a-metrics-client-ca\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.953038 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.952920 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-accelerators-collector-config\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.957436 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.956322 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.957436 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.957341 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/d148eabc-a2af-41d8-ba7d-039b9da0132a-node-exporter-tls\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.959902 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.959869 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqgvc\" (UniqueName: \"kubernetes.io/projected/d148eabc-a2af-41d8-ba7d-039b9da0132a-kube-api-access-pqgvc\") pod \"node-exporter-bzjx9\" (UID: \"d148eabc-a2af-41d8-ba7d-039b9da0132a\") " pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:08.988652 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:08.988562 2574 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1bd8c4c-012e-4776-934f-9a95a15ce4bb" path="/var/lib/kubelet/pods/d1bd8c4c-012e-4776-934f-9a95a15ce4bb/volumes" Apr 22 14:18:09.026112 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.026084 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" Apr 22 14:18:09.038454 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.038430 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8"] Apr 22 14:18:09.040830 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.040804 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-bzjx9" Apr 22 14:18:09.040927 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:18:09.040872 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82e2b9ba_996b_4f92_b967_0052faa65d73.slice/crio-dd923a4b66e23548fe23265e3745d8f76637d04310c2d4e4a3643a5cdd269a4d WatchSource:0}: Error finding container dd923a4b66e23548fe23265e3745d8f76637d04310c2d4e4a3643a5cdd269a4d: Status 404 returned error can't find the container with id dd923a4b66e23548fe23265e3745d8f76637d04310c2d4e4a3643a5cdd269a4d Apr 22 14:18:09.050723 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:18:09.050661 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd148eabc_a2af_41d8_ba7d_039b9da0132a.slice/crio-859eb27645948e6dc933694fb1ba33529557ce5f693e86967cf3cbd6e6f16ed8 WatchSource:0}: Error finding container 859eb27645948e6dc933694fb1ba33529557ce5f693e86967cf3cbd6e6f16ed8: Status 404 returned error can't find the container with id 859eb27645948e6dc933694fb1ba33529557ce5f693e86967cf3cbd6e6f16ed8 Apr 22 14:18:09.165732 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.165661 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-69db897b98-sjd66"] Apr 22 14:18:09.168443 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:18:09.168418 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeeda31a_52ab_4389_b06c_eea0f137d09f.slice/crio-0ddca99b80aa7b1f9a0b1bf3f007ba2784df054eee2a79d71002b0517db4af20 WatchSource:0}: Error finding container 0ddca99b80aa7b1f9a0b1bf3f007ba2784df054eee2a79d71002b0517db4af20: Status 404 returned error can't find the container with id 0ddca99b80aa7b1f9a0b1bf3f007ba2784df054eee2a79d71002b0517db4af20 Apr 22 14:18:09.536360 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.536203 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" event={"ID":"82e2b9ba-996b-4f92-b967-0052faa65d73","Type":"ContainerStarted","Data":"3c07e5b35f9b693b3c707e46265c13838bb95d20297cff64820bede906aef1a5"} Apr 22 14:18:09.536360 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.536247 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" event={"ID":"82e2b9ba-996b-4f92-b967-0052faa65d73","Type":"ContainerStarted","Data":"e28455be80481e2e734e121e271fb539fa680b755f3db0b3471086b031c965ef"} Apr 22 14:18:09.536360 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.536263 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" event={"ID":"82e2b9ba-996b-4f92-b967-0052faa65d73","Type":"ContainerStarted","Data":"dd923a4b66e23548fe23265e3745d8f76637d04310c2d4e4a3643a5cdd269a4d"} Apr 22 14:18:09.537479 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.537448 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bzjx9" event={"ID":"d148eabc-a2af-41d8-ba7d-039b9da0132a","Type":"ContainerStarted","Data":"859eb27645948e6dc933694fb1ba33529557ce5f693e86967cf3cbd6e6f16ed8"} Apr 22 14:18:09.538588 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.538543 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" event={"ID":"eeeda31a-52ab-4389-b06c-eea0f137d09f","Type":"ContainerStarted","Data":"0ddca99b80aa7b1f9a0b1bf3f007ba2784df054eee2a79d71002b0517db4af20"} Apr 22 14:18:09.653503 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.653414 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 14:18:09.657393 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.657363 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.659999 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.659734 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Apr 22 14:18:09.659999 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.659822 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Apr 22 14:18:09.659999 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.659869 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Apr 22 14:18:09.661231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.660752 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Apr 22 14:18:09.661231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.660815 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Apr 22 14:18:09.661231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.660899 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-26xrv\"" Apr 22 14:18:09.661231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.660953 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Apr 22 14:18:09.661231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.661002 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Apr 22 14:18:09.661231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.661056 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Apr 22 14:18:09.661231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.661096 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Apr 22 14:18:09.676774 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.676741 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 14:18:09.763095 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763056 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-web-config\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763244 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763114 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l99wd\" (UniqueName: \"kubernetes.io/projected/19482d30-2da5-4d8a-89da-515e692e97c7-kube-api-access-l99wd\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763244 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763145 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/19482d30-2da5-4d8a-89da-515e692e97c7-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763244 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763180 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763244 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763209 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/19482d30-2da5-4d8a-89da-515e692e97c7-config-out\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763244 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763233 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19482d30-2da5-4d8a-89da-515e692e97c7-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763492 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763264 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763492 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763312 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/19482d30-2da5-4d8a-89da-515e692e97c7-tls-assets\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763492 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763473 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/19482d30-2da5-4d8a-89da-515e692e97c7-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763627 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763514 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763627 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763562 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-config-volume\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763627 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763595 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.763797 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.763665 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.864571 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864484 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-web-config\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.864571 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864524 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l99wd\" (UniqueName: \"kubernetes.io/projected/19482d30-2da5-4d8a-89da-515e692e97c7-kube-api-access-l99wd\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.864839 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864629 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/19482d30-2da5-4d8a-89da-515e692e97c7-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.864839 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864713 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.864839 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864750 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/19482d30-2da5-4d8a-89da-515e692e97c7-config-out\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.864839 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864779 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19482d30-2da5-4d8a-89da-515e692e97c7-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.864839 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864828 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.865093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864876 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/19482d30-2da5-4d8a-89da-515e692e97c7-tls-assets\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.865093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.864951 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/19482d30-2da5-4d8a-89da-515e692e97c7-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.865859 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.865833 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/19482d30-2da5-4d8a-89da-515e692e97c7-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.865859 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.865843 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19482d30-2da5-4d8a-89da-515e692e97c7-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.866078 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.865897 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.866078 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.865950 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-config-volume\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.866282 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.866255 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.866362 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.866338 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.866972 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.866948 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/19482d30-2da5-4d8a-89da-515e692e97c7-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.868008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.867958 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.868637 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.868482 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-web-config\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.868897 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.868852 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.869137 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.869095 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/19482d30-2da5-4d8a-89da-515e692e97c7-tls-assets\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.869571 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.869526 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-config-volume\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.870217 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.870175 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.870739 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.870704 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.871399 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.871171 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/19482d30-2da5-4d8a-89da-515e692e97c7-config-out\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.871489 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.871473 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/19482d30-2da5-4d8a-89da-515e692e97c7-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.873728 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.873665 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l99wd\" (UniqueName: \"kubernetes.io/projected/19482d30-2da5-4d8a-89da-515e692e97c7-kube-api-access-l99wd\") pod \"alertmanager-main-0\" (UID: \"19482d30-2da5-4d8a-89da-515e692e97c7\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:09.971347 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:09.971278 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 14:18:10.645286 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:10.645265 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 14:18:10.648185 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:18:10.648142 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19482d30_2da5_4d8a_89da_515e692e97c7.slice/crio-e55353c914b463e4803979d887d247174b225a9d27f4f1dd5b9bd60ea5b14f9a WatchSource:0}: Error finding container e55353c914b463e4803979d887d247174b225a9d27f4f1dd5b9bd60ea5b14f9a: Status 404 returned error can't find the container with id e55353c914b463e4803979d887d247174b225a9d27f4f1dd5b9bd60ea5b14f9a Apr 22 14:18:11.547359 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.547299 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" event={"ID":"82e2b9ba-996b-4f92-b967-0052faa65d73","Type":"ContainerStarted","Data":"2a31146514eb5913e199843248c95263f75c61cfd57d9b31d069a6b4211a2fca"} Apr 22 14:18:11.549304 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.549274 2574 generic.go:358] "Generic (PLEG): container finished" podID="d148eabc-a2af-41d8-ba7d-039b9da0132a" containerID="b7f3a64590121602c7a6f3b52c93336169df0841fe6f62ca5c017ce83d43c275" exitCode=0 Apr 22 14:18:11.549449 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.549329 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bzjx9" event={"ID":"d148eabc-a2af-41d8-ba7d-039b9da0132a","Type":"ContainerDied","Data":"b7f3a64590121602c7a6f3b52c93336169df0841fe6f62ca5c017ce83d43c275"} Apr 22 14:18:11.551826 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.551796 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" event={"ID":"eeeda31a-52ab-4389-b06c-eea0f137d09f","Type":"ContainerStarted","Data":"af69ce2b132dc9fcf827676555dce3cad5a38a52ebfbab8333d2e189e69d60c6"} Apr 22 14:18:11.552046 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.551833 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" event={"ID":"eeeda31a-52ab-4389-b06c-eea0f137d09f","Type":"ContainerStarted","Data":"773af581c898d902d2e9f5f7387ed7d9c102754ded19375b19dba24712479514"} Apr 22 14:18:11.552046 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.551847 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" event={"ID":"eeeda31a-52ab-4389-b06c-eea0f137d09f","Type":"ContainerStarted","Data":"415570d0f65faf09166f348a8fa1f73aade34902b16847458912b1a566ed0312"} Apr 22 14:18:11.553482 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.553435 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerStarted","Data":"e55353c914b463e4803979d887d247174b225a9d27f4f1dd5b9bd60ea5b14f9a"} Apr 22 14:18:11.570386 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.570343 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-lvzv8" podStartSLOduration=2.260310654 podStartE2EDuration="3.570324178s" podCreationTimestamp="2026-04-22 14:18:08 +0000 UTC" firstStartedPulling="2026-04-22 14:18:09.171882865 +0000 UTC m=+202.639760573" lastFinishedPulling="2026-04-22 14:18:10.481896385 +0000 UTC m=+203.949774097" observedRunningTime="2026-04-22 14:18:11.56841413 +0000 UTC m=+205.036291860" watchObservedRunningTime="2026-04-22 14:18:11.570324178 +0000 UTC m=+205.038201909" Apr 22 14:18:11.619708 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:11.619350 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-69db897b98-sjd66" podStartSLOduration=2.303493014 podStartE2EDuration="3.619329592s" podCreationTimestamp="2026-04-22 14:18:08 +0000 UTC" firstStartedPulling="2026-04-22 14:18:09.170500937 +0000 UTC m=+202.638378646" lastFinishedPulling="2026-04-22 14:18:10.486337502 +0000 UTC m=+203.954215224" observedRunningTime="2026-04-22 14:18:11.617596747 +0000 UTC m=+205.085474490" watchObservedRunningTime="2026-04-22 14:18:11.619329592 +0000 UTC m=+205.087207323" Apr 22 14:18:12.558853 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:12.558821 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bzjx9" event={"ID":"d148eabc-a2af-41d8-ba7d-039b9da0132a","Type":"ContainerStarted","Data":"0c1e41fdaea79e083157dfba9a2fc93f1d27f2cf1e23316cdb2c01ee739e8be9"} Apr 22 14:18:12.559300 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:12.558861 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bzjx9" event={"ID":"d148eabc-a2af-41d8-ba7d-039b9da0132a","Type":"ContainerStarted","Data":"d506b5a9b63239907cffa8a9cec65da3837ac4d043c9698d54aa21d284e50426"} Apr 22 14:18:12.560166 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:12.560141 2574 generic.go:358] "Generic (PLEG): container finished" podID="19482d30-2da5-4d8a-89da-515e692e97c7" containerID="d3d45f3146f90a3a09da926c0066f6dcfea72d2516042f1bcf33cdaf41c10a9f" exitCode=0 Apr 22 14:18:12.560260 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:12.560218 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerDied","Data":"d3d45f3146f90a3a09da926c0066f6dcfea72d2516042f1bcf33cdaf41c10a9f"} Apr 22 14:18:12.579992 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:12.579947 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-bzjx9" podStartSLOduration=3.150500961 podStartE2EDuration="4.57992085s" podCreationTimestamp="2026-04-22 14:18:08 +0000 UTC" firstStartedPulling="2026-04-22 14:18:09.052489747 +0000 UTC m=+202.520367461" lastFinishedPulling="2026-04-22 14:18:10.481909635 +0000 UTC m=+203.949787350" observedRunningTime="2026-04-22 14:18:12.57782048 +0000 UTC m=+206.045698210" watchObservedRunningTime="2026-04-22 14:18:12.57992085 +0000 UTC m=+206.047798579" Apr 22 14:18:13.046380 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.046350 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-b95c9c57c-c2st2"] Apr 22 14:18:13.049727 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.049702 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.052404 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.052383 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-server-audit-profiles\"" Apr 22 14:18:13.053550 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.053522 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-tls\"" Apr 22 14:18:13.053697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.053595 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-dockercfg-rdhsm\"" Apr 22 14:18:13.053697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.053616 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-client-certs\"" Apr 22 14:18:13.053697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.053634 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kubelet-serving-ca-bundle\"" Apr 22 14:18:13.055511 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.055493 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-dlq1o7b9kiamq\"" Apr 22 14:18:13.060415 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.060393 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-b95c9c57c-c2st2"] Apr 22 14:18:13.099974 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.099946 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.100096 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.100029 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-metrics-server-audit-profiles\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.100158 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.100114 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-secret-metrics-server-client-certs\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.100207 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.100175 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-secret-metrics-server-tls\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.100240 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.100203 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-client-ca-bundle\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.100299 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.100282 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-audit-log\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.100346 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.100332 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc474\" (UniqueName: \"kubernetes.io/projected/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-kube-api-access-qc474\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.201596 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.201562 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.201827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.201635 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-metrics-server-audit-profiles\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.201827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.201691 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-secret-metrics-server-client-certs\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.201946 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.201908 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-secret-metrics-server-tls\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.201997 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.201969 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-client-ca-bundle\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.203648 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.202724 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.203648 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.202879 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-audit-log\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.203648 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.202977 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qc474\" (UniqueName: \"kubernetes.io/projected/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-kube-api-access-qc474\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.203648 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.203371 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-audit-log\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.204003 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.203895 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-metrics-server-audit-profiles\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.204903 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.204873 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-secret-metrics-server-client-certs\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.205918 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.205896 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-secret-metrics-server-tls\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.206729 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.206713 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-client-ca-bundle\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.227556 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.227528 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc474\" (UniqueName: \"kubernetes.io/projected/2e68eca2-c4cf-4ccf-994b-69c0a0b47de4-kube-api-access-qc474\") pod \"metrics-server-b95c9c57c-c2st2\" (UID: \"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4\") " pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.363178 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.363092 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:13.529503 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:13.529467 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-b95c9c57c-c2st2"] Apr 22 14:18:13.847885 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:18:13.847835 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e68eca2_c4cf_4ccf_994b_69c0a0b47de4.slice/crio-1c44033bc33dd3883d6fe654c4ba1e1a8883f08509f7a253335db0dde9d7edf2 WatchSource:0}: Error finding container 1c44033bc33dd3883d6fe654c4ba1e1a8883f08509f7a253335db0dde9d7edf2: Status 404 returned error can't find the container with id 1c44033bc33dd3883d6fe654c4ba1e1a8883f08509f7a253335db0dde9d7edf2 Apr 22 14:18:14.570785 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.570735 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerStarted","Data":"2f8d44e22fd7efe408ce3450915ea8187f559588759a40fa197d010ffb86a2c8"} Apr 22 14:18:14.570785 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.570783 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerStarted","Data":"1bba5fc090bf561ef9a2c91bee9e5771eb7b4d6b55613fc971fd99b00e10d635"} Apr 22 14:18:14.571057 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.570799 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerStarted","Data":"62bd156162e1ca38f1bff48cff99421e800b56e325899493dcde33b396b46c95"} Apr 22 14:18:14.571057 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.570812 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerStarted","Data":"997048f240974aed0b2b0ff66fd54887658ef340cee58e6d2962bb67f59f83f7"} Apr 22 14:18:14.571057 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.570825 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerStarted","Data":"ce5152998da9e735d5d429e6514bc1772181d4f21d27d235a9308e7cb239a0f8"} Apr 22 14:18:14.573572 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.573192 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" event={"ID":"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4","Type":"ContainerStarted","Data":"1c44033bc33dd3883d6fe654c4ba1e1a8883f08509f7a253335db0dde9d7edf2"} Apr 22 14:18:14.875890 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.875818 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 14:18:14.879939 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.879916 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:14.883598 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.883566 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Apr 22 14:18:14.883843 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.883821 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Apr 22 14:18:14.883948 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.883910 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Apr 22 14:18:14.884023 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.883978 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-an66lmohunusd\"" Apr 22 14:18:14.884079 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.884040 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-42c4h\"" Apr 22 14:18:14.884232 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.884211 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Apr 22 14:18:14.884430 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.884402 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Apr 22 14:18:14.884430 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.884394 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Apr 22 14:18:14.884614 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.884598 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Apr 22 14:18:14.884926 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.884906 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Apr 22 14:18:14.885029 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.884983 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Apr 22 14:18:14.886529 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.886419 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Apr 22 14:18:14.886529 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.886436 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Apr 22 14:18:14.892264 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.892248 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Apr 22 14:18:14.959880 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:14.959840 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 14:18:15.019696 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.019638 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/114db3c0-b406-4a1d-b927-efb1d4fd0e66-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.019857 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.019719 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.019857 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.019766 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/114db3c0-b406-4a1d-b927-efb1d4fd0e66-config-out\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.019946 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.019877 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.019983 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.019960 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020034 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.019988 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020034 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020016 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020091 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020042 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn7cd\" (UniqueName: \"kubernetes.io/projected/114db3c0-b406-4a1d-b927-efb1d4fd0e66-kube-api-access-qn7cd\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020137 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020107 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020142 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020201 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020163 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020271 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020202 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-config\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020271 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020236 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020352 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020268 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020352 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020307 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020352 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020324 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020464 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020370 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-web-config\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.020464 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.020442 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.121525 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121487 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.121748 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121545 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/114db3c0-b406-4a1d-b927-efb1d4fd0e66-config-out\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.121748 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121597 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.121748 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121660 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.121748 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121714 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.121748 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121746 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.122012 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121772 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qn7cd\" (UniqueName: \"kubernetes.io/projected/114db3c0-b406-4a1d-b927-efb1d4fd0e66-kube-api-access-qn7cd\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.122012 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.121837 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.122615 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.122534 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.122615 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.122579 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.122615 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.122616 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-config\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.122936 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.122664 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123367 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123416 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123459 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123475 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123491 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123541 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-web-config\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123586 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.123836 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.123625 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/114db3c0-b406-4a1d-b927-efb1d4fd0e66-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.125423 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.124975 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/114db3c0-b406-4a1d-b927-efb1d4fd0e66-config-out\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.125423 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.125122 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.125423 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.125375 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.126340 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.126275 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.126806 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.126731 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.126899 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.126850 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.126972 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.126929 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/114db3c0-b406-4a1d-b927-efb1d4fd0e66-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.127199 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.127164 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.127557 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.127534 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-web-config\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.127903 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.127880 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/114db3c0-b406-4a1d-b927-efb1d4fd0e66-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.128157 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.128132 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.128609 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.128584 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.128787 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.128769 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-config\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.129063 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.129009 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.129893 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.129871 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/114db3c0-b406-4a1d-b927-efb1d4fd0e66-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.131137 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.131114 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn7cd\" (UniqueName: \"kubernetes.io/projected/114db3c0-b406-4a1d-b927-efb1d4fd0e66-kube-api-access-qn7cd\") pod \"prometheus-k8s-0\" (UID: \"114db3c0-b406-4a1d-b927-efb1d4fd0e66\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.193129 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.193096 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:15.399487 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.399463 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 14:18:15.401596 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:18:15.401569 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod114db3c0_b406_4a1d_b927_efb1d4fd0e66.slice/crio-e6aef7e1d673a6d987287a22837bc5b04f12b88e8dabce793effb099ae5a9e90 WatchSource:0}: Error finding container e6aef7e1d673a6d987287a22837bc5b04f12b88e8dabce793effb099ae5a9e90: Status 404 returned error can't find the container with id e6aef7e1d673a6d987287a22837bc5b04f12b88e8dabce793effb099ae5a9e90 Apr 22 14:18:15.579908 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.579868 2574 generic.go:358] "Generic (PLEG): container finished" podID="114db3c0-b406-4a1d-b927-efb1d4fd0e66" containerID="50447878a5eb97612718d507d65b1e63f7cc4f0b9620f4dce9be2f7354f4c1dd" exitCode=0 Apr 22 14:18:15.580044 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.579930 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerDied","Data":"50447878a5eb97612718d507d65b1e63f7cc4f0b9620f4dce9be2f7354f4c1dd"} Apr 22 14:18:15.580044 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.579971 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerStarted","Data":"e6aef7e1d673a6d987287a22837bc5b04f12b88e8dabce793effb099ae5a9e90"} Apr 22 14:18:15.582988 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.582962 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"19482d30-2da5-4d8a-89da-515e692e97c7","Type":"ContainerStarted","Data":"f7e3271c812d772d6dd2068975e1f44eade290a1df6f5bd243e2cb8a0018909c"} Apr 22 14:18:15.584228 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.584210 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" event={"ID":"2e68eca2-c4cf-4ccf-994b-69c0a0b47de4","Type":"ContainerStarted","Data":"3fe1a813378d554a399ba1195b8f6dcebb896ecc8e664afd359168398a879865"} Apr 22 14:18:15.642039 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.641978 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.037682685 podStartE2EDuration="6.641961905s" podCreationTimestamp="2026-04-22 14:18:09 +0000 UTC" firstStartedPulling="2026-04-22 14:18:10.650975131 +0000 UTC m=+204.118852839" lastFinishedPulling="2026-04-22 14:18:15.255254332 +0000 UTC m=+208.723132059" observedRunningTime="2026-04-22 14:18:15.641149787 +0000 UTC m=+209.109027517" watchObservedRunningTime="2026-04-22 14:18:15.641961905 +0000 UTC m=+209.109839637" Apr 22 14:18:15.661114 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:15.661031 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" podStartSLOduration=1.252639174 podStartE2EDuration="2.661017365s" podCreationTimestamp="2026-04-22 14:18:13 +0000 UTC" firstStartedPulling="2026-04-22 14:18:13.849811329 +0000 UTC m=+207.317689038" lastFinishedPulling="2026-04-22 14:18:15.258189515 +0000 UTC m=+208.726067229" observedRunningTime="2026-04-22 14:18:15.659487518 +0000 UTC m=+209.127365261" watchObservedRunningTime="2026-04-22 14:18:15.661017365 +0000 UTC m=+209.128895150" Apr 22 14:18:18.595202 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:18.595173 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerStarted","Data":"06380b09e55d1362d50fe6366e6e3cacaa588526696899a328e0dd6c04e60bae"} Apr 22 14:18:19.600354 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:19.600319 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerStarted","Data":"24bf89ce35f66bd519ad466bda5267c0e5e742654a3c4114f9ed4a278723f41c"} Apr 22 14:18:20.606058 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:20.606030 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerStarted","Data":"5d4881648b779cc4412760290488541e36e0750698864eb78cf54535cb36bcc6"} Apr 22 14:18:20.606381 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:20.606071 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerStarted","Data":"72ce776c8bea34b904804d590be9a5f046c993ac9606c37071052819b7ac60b4"} Apr 22 14:18:20.606381 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:20.606085 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerStarted","Data":"9c3f8b17b16a01aa967f7bc4594b2030cd16f2183103175e78d04e7a14525355"} Apr 22 14:18:21.612636 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:21.612602 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"114db3c0-b406-4a1d-b927-efb1d4fd0e66","Type":"ContainerStarted","Data":"96d3dd817987032c1fb60f0e1e6216382e36f689cbd963ecc16751a8ecaa3ad6"} Apr 22 14:18:21.640570 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:21.640518 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=2.872916826 podStartE2EDuration="7.64049948s" podCreationTimestamp="2026-04-22 14:18:14 +0000 UTC" firstStartedPulling="2026-04-22 14:18:15.58116573 +0000 UTC m=+209.049043438" lastFinishedPulling="2026-04-22 14:18:20.34874837 +0000 UTC m=+213.816626092" observedRunningTime="2026-04-22 14:18:21.638940605 +0000 UTC m=+215.106818352" watchObservedRunningTime="2026-04-22 14:18:21.64049948 +0000 UTC m=+215.108377210" Apr 22 14:18:25.193825 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:25.193761 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:18:27.607755 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.607711 2574 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-588c45675c-6nbl6" podUID="b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" containerName="console" containerID="cri-o://6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e" gracePeriod=15 Apr 22 14:18:27.842796 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.842774 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-588c45675c-6nbl6_b79bdabc-7a55-43ee-a856-0cbc5b1c49c6/console/0.log" Apr 22 14:18:27.842913 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.842831 2574 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:18:27.849356 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849339 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-serving-cert\") pod \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " Apr 22 14:18:27.849412 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849400 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-service-ca\") pod \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " Apr 22 14:18:27.849519 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849498 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-config\") pod \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " Apr 22 14:18:27.849562 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849545 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-oauth-config\") pod \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " Apr 22 14:18:27.849597 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849564 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtsfs\" (UniqueName: \"kubernetes.io/projected/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-kube-api-access-wtsfs\") pod \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " Apr 22 14:18:27.849636 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849614 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-oauth-serving-cert\") pod \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\" (UID: \"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6\") " Apr 22 14:18:27.849854 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849813 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-service-ca" (OuterVolumeSpecName: "service-ca") pod "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" (UID: "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:27.849997 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849978 2574 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-service-ca\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:27.849997 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.849985 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-config" (OuterVolumeSpecName: "console-config") pod "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" (UID: "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:27.850127 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.850092 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" (UID: "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:27.851878 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.851855 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-kube-api-access-wtsfs" (OuterVolumeSpecName: "kube-api-access-wtsfs") pod "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" (UID: "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6"). InnerVolumeSpecName "kube-api-access-wtsfs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 14:18:27.851962 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.851877 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" (UID: "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 14:18:27.851962 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.851888 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" (UID: "b79bdabc-7a55-43ee-a856-0cbc5b1c49c6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 14:18:27.951270 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.951234 2574 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-config\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:27.951270 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.951266 2574 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-oauth-config\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:27.951270 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.951277 2574 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wtsfs\" (UniqueName: \"kubernetes.io/projected/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-kube-api-access-wtsfs\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:27.951497 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.951286 2574 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-oauth-serving-cert\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:27.951497 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:27.951296 2574 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6-console-serving-cert\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:28.635136 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.635110 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-588c45675c-6nbl6_b79bdabc-7a55-43ee-a856-0cbc5b1c49c6/console/0.log" Apr 22 14:18:28.635512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.635149 2574 generic.go:358] "Generic (PLEG): container finished" podID="b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" containerID="6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e" exitCode=2 Apr 22 14:18:28.635512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.635218 2574 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-588c45675c-6nbl6" Apr 22 14:18:28.635512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.635216 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-588c45675c-6nbl6" event={"ID":"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6","Type":"ContainerDied","Data":"6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e"} Apr 22 14:18:28.635512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.635324 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-588c45675c-6nbl6" event={"ID":"b79bdabc-7a55-43ee-a856-0cbc5b1c49c6","Type":"ContainerDied","Data":"7f295c412dedc31c663ed330c2e8f8e0f80315ea36a5c2f9fa7dce39568520d8"} Apr 22 14:18:28.635512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.635344 2574 scope.go:117] "RemoveContainer" containerID="6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e" Apr 22 14:18:28.643733 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.643712 2574 scope.go:117] "RemoveContainer" containerID="6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e" Apr 22 14:18:28.644004 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:18:28.643985 2574 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e\": container with ID starting with 6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e not found: ID does not exist" containerID="6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e" Apr 22 14:18:28.644056 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.644017 2574 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e"} err="failed to get container status \"6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e\": rpc error: code = NotFound desc = could not find container \"6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e\": container with ID starting with 6307fe2dc486cf554f82ceffe9efb126376c94c1e05e561498b5295ea6bf538e not found: ID does not exist" Apr 22 14:18:28.660194 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.660167 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-588c45675c-6nbl6"] Apr 22 14:18:28.664900 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.664877 2574 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-588c45675c-6nbl6"] Apr 22 14:18:28.730532 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.730502 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7df65c58c4-27ft2"] Apr 22 14:18:28.987927 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:28.987796 2574 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" path="/var/lib/kubelet/pods/b79bdabc-7a55-43ee-a856-0cbc5b1c49c6/volumes" Apr 22 14:18:32.647600 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:32.647570 2574 generic.go:358] "Generic (PLEG): container finished" podID="36bf42f9-41dd-4d29-9f4c-2472e7df5572" containerID="40ed85ed8d94878671a089ee4a421e20b969b67d97186696df3bf56a3a1a5e65" exitCode=0 Apr 22 14:18:32.647981 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:32.647623 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" event={"ID":"36bf42f9-41dd-4d29-9f4c-2472e7df5572","Type":"ContainerDied","Data":"40ed85ed8d94878671a089ee4a421e20b969b67d97186696df3bf56a3a1a5e65"} Apr 22 14:18:32.647981 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:32.647960 2574 scope.go:117] "RemoveContainer" containerID="40ed85ed8d94878671a089ee4a421e20b969b67d97186696df3bf56a3a1a5e65" Apr 22 14:18:33.364047 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:33.364010 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:33.364047 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:33.364053 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:33.652598 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:33.652562 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-9lcw9" event={"ID":"36bf42f9-41dd-4d29-9f4c-2472e7df5572","Type":"ContainerStarted","Data":"63324b1df916300ab348bf59b8bb866254e4f43590b97f72d25abbb11db52a40"} Apr 22 14:18:47.697988 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:47.697888 2574 generic.go:358] "Generic (PLEG): container finished" podID="727af150-86df-4722-a839-797b4cb1db69" containerID="e6bc2f7747be75ad033a180b6120f4d20088c3ec0eac917bfa1b9a3c991da0cb" exitCode=0 Apr 22 14:18:47.697988 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:47.697963 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-zkfsf" event={"ID":"727af150-86df-4722-a839-797b4cb1db69","Type":"ContainerDied","Data":"e6bc2f7747be75ad033a180b6120f4d20088c3ec0eac917bfa1b9a3c991da0cb"} Apr 22 14:18:47.698542 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:47.698403 2574 scope.go:117] "RemoveContainer" containerID="e6bc2f7747be75ad033a180b6120f4d20088c3ec0eac917bfa1b9a3c991da0cb" Apr 22 14:18:48.702742 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:48.702704 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-zkfsf" event={"ID":"727af150-86df-4722-a839-797b4cb1db69","Type":"ContainerStarted","Data":"76240bcfb752bd5d2a02a359e80023209a3d77e0664ac289d358d6ae1f2ae7b0"} Apr 22 14:18:53.369265 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:53.369236 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:53.373272 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:53.373247 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-b95c9c57c-c2st2" Apr 22 14:18:53.750555 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:53.750502 2574 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-7df65c58c4-27ft2" podUID="c6863efe-089d-4c10-a59b-7cc11a649cc1" containerName="console" containerID="cri-o://b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e" gracePeriod=15 Apr 22 14:18:54.022556 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.022531 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7df65c58c4-27ft2_c6863efe-089d-4c10-a59b-7cc11a649cc1/console/0.log" Apr 22 14:18:54.022730 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.022605 2574 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:18:54.093308 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093272 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhvtd\" (UniqueName: \"kubernetes.io/projected/c6863efe-089d-4c10-a59b-7cc11a649cc1-kube-api-access-fhvtd\") pod \"c6863efe-089d-4c10-a59b-7cc11a649cc1\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " Apr 22 14:18:54.093475 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093332 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-trusted-ca-bundle\") pod \"c6863efe-089d-4c10-a59b-7cc11a649cc1\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " Apr 22 14:18:54.093475 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093363 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-config\") pod \"c6863efe-089d-4c10-a59b-7cc11a649cc1\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " Apr 22 14:18:54.093475 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093401 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-service-ca\") pod \"c6863efe-089d-4c10-a59b-7cc11a649cc1\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " Apr 22 14:18:54.093637 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093534 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-oauth-config\") pod \"c6863efe-089d-4c10-a59b-7cc11a649cc1\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " Apr 22 14:18:54.093716 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093639 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-oauth-serving-cert\") pod \"c6863efe-089d-4c10-a59b-7cc11a649cc1\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " Apr 22 14:18:54.093837 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093813 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-config" (OuterVolumeSpecName: "console-config") pod "c6863efe-089d-4c10-a59b-7cc11a649cc1" (UID: "c6863efe-089d-4c10-a59b-7cc11a649cc1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:54.093896 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093819 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c6863efe-089d-4c10-a59b-7cc11a649cc1" (UID: "c6863efe-089d-4c10-a59b-7cc11a649cc1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:54.093962 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.093934 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-service-ca" (OuterVolumeSpecName: "service-ca") pod "c6863efe-089d-4c10-a59b-7cc11a649cc1" (UID: "c6863efe-089d-4c10-a59b-7cc11a649cc1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:54.094209 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.094029 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-serving-cert\") pod \"c6863efe-089d-4c10-a59b-7cc11a649cc1\" (UID: \"c6863efe-089d-4c10-a59b-7cc11a649cc1\") " Apr 22 14:18:54.094292 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.094207 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c6863efe-089d-4c10-a59b-7cc11a649cc1" (UID: "c6863efe-089d-4c10-a59b-7cc11a649cc1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:18:54.094967 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.094453 2574 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-service-ca\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:54.094967 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.094475 2574 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-oauth-serving-cert\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:54.094967 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.094489 2574 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-trusted-ca-bundle\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:54.094967 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.094505 2574 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-config\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:54.096390 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.096362 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c6863efe-089d-4c10-a59b-7cc11a649cc1" (UID: "c6863efe-089d-4c10-a59b-7cc11a649cc1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 14:18:54.096509 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.096362 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6863efe-089d-4c10-a59b-7cc11a649cc1-kube-api-access-fhvtd" (OuterVolumeSpecName: "kube-api-access-fhvtd") pod "c6863efe-089d-4c10-a59b-7cc11a649cc1" (UID: "c6863efe-089d-4c10-a59b-7cc11a649cc1"). InnerVolumeSpecName "kube-api-access-fhvtd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 14:18:54.096770 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.096741 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c6863efe-089d-4c10-a59b-7cc11a649cc1" (UID: "c6863efe-089d-4c10-a59b-7cc11a649cc1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 14:18:54.195369 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.195336 2574 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-serving-cert\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:54.195369 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.195367 2574 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-fhvtd\" (UniqueName: \"kubernetes.io/projected/c6863efe-089d-4c10-a59b-7cc11a649cc1-kube-api-access-fhvtd\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:54.195567 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.195379 2574 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c6863efe-089d-4c10-a59b-7cc11a649cc1-console-oauth-config\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:18:54.724784 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.724754 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7df65c58c4-27ft2_c6863efe-089d-4c10-a59b-7cc11a649cc1/console/0.log" Apr 22 14:18:54.725225 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.724797 2574 generic.go:358] "Generic (PLEG): container finished" podID="c6863efe-089d-4c10-a59b-7cc11a649cc1" containerID="b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e" exitCode=2 Apr 22 14:18:54.725225 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.724844 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7df65c58c4-27ft2" event={"ID":"c6863efe-089d-4c10-a59b-7cc11a649cc1","Type":"ContainerDied","Data":"b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e"} Apr 22 14:18:54.725225 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.724862 2574 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7df65c58c4-27ft2" Apr 22 14:18:54.725225 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.724872 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7df65c58c4-27ft2" event={"ID":"c6863efe-089d-4c10-a59b-7cc11a649cc1","Type":"ContainerDied","Data":"7310873f1ae2711bed8313cb923742aaf959c1034a5c21cc27e67ae7855e5d69"} Apr 22 14:18:54.725225 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.724894 2574 scope.go:117] "RemoveContainer" containerID="b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e" Apr 22 14:18:54.733354 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.733331 2574 scope.go:117] "RemoveContainer" containerID="b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e" Apr 22 14:18:54.733626 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:18:54.733607 2574 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e\": container with ID starting with b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e not found: ID does not exist" containerID="b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e" Apr 22 14:18:54.733702 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.733635 2574 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e"} err="failed to get container status \"b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e\": rpc error: code = NotFound desc = could not find container \"b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e\": container with ID starting with b8cc6eccae44fc086e65140db023731243ae6df127546dae06effc8b35986e6e not found: ID does not exist" Apr 22 14:18:54.745093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.745055 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7df65c58c4-27ft2"] Apr 22 14:18:54.751506 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.751478 2574 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-7df65c58c4-27ft2"] Apr 22 14:18:54.987374 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:54.987292 2574 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6863efe-089d-4c10-a59b-7cc11a649cc1" path="/var/lib/kubelet/pods/c6863efe-089d-4c10-a59b-7cc11a649cc1/volumes" Apr 22 14:18:58.839558 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:58.839522 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:18:58.842212 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:58.842178 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1eca98e2-015f-4ce7-bcd6-76878fa26d4b-metrics-certs\") pod \"network-metrics-daemon-p65qg\" (UID: \"1eca98e2-015f-4ce7-bcd6-76878fa26d4b\") " pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:18:59.087368 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:59.087331 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-4nhsh\"" Apr 22 14:18:59.095220 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:59.095133 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p65qg" Apr 22 14:18:59.227456 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:59.227427 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p65qg"] Apr 22 14:18:59.230275 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:18:59.230227 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1eca98e2_015f_4ce7_bcd6_76878fa26d4b.slice/crio-a6433cc51ffc8df34a4510409ec32d112c493448626f530fc8121ff6aeab7f0e WatchSource:0}: Error finding container a6433cc51ffc8df34a4510409ec32d112c493448626f530fc8121ff6aeab7f0e: Status 404 returned error can't find the container with id a6433cc51ffc8df34a4510409ec32d112c493448626f530fc8121ff6aeab7f0e Apr 22 14:18:59.743185 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:18:59.743150 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p65qg" event={"ID":"1eca98e2-015f-4ce7-bcd6-76878fa26d4b","Type":"ContainerStarted","Data":"a6433cc51ffc8df34a4510409ec32d112c493448626f530fc8121ff6aeab7f0e"} Apr 22 14:19:00.755488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:00.755454 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p65qg" event={"ID":"1eca98e2-015f-4ce7-bcd6-76878fa26d4b","Type":"ContainerStarted","Data":"8c9fd05345bd76915897dd663583974431085e85cb47770a98903a0a7bb154d4"} Apr 22 14:19:00.755488 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:00.755492 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p65qg" event={"ID":"1eca98e2-015f-4ce7-bcd6-76878fa26d4b","Type":"ContainerStarted","Data":"2646c805c149ab36515a38e51ead694c47582e0b848f119be6e70028ce8d5f47"} Apr 22 14:19:00.777466 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:00.777405 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-p65qg" podStartSLOduration=252.65289576 podStartE2EDuration="4m13.777385503s" podCreationTimestamp="2026-04-22 14:14:47 +0000 UTC" firstStartedPulling="2026-04-22 14:18:59.232199556 +0000 UTC m=+252.700077265" lastFinishedPulling="2026-04-22 14:19:00.356689287 +0000 UTC m=+253.824567008" observedRunningTime="2026-04-22 14:19:00.776448958 +0000 UTC m=+254.244326689" watchObservedRunningTime="2026-04-22 14:19:00.777385503 +0000 UTC m=+254.245263234" Apr 22 14:19:15.194117 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:15.194022 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:19:15.213373 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:15.213344 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:19:15.818602 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:15.818570 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 14:19:26.382344 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:19:26.382297 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-b6brs" podUID="daaf870a-47fd-4fba-8f87-375c235c1a4c" Apr 22 14:19:26.382344 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:19:26.382297 2574 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-4kpkf" podUID="0ef3c015-11a0-43bf-b3db-a4ee755ab04b" Apr 22 14:19:26.836694 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:26.836634 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4kpkf" Apr 22 14:19:26.836848 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:26.836720 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:19:29.519752 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.519713 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:19:29.520175 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.519802 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:19:29.522360 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.522321 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0ef3c015-11a0-43bf-b3db-a4ee755ab04b-metrics-tls\") pod \"dns-default-4kpkf\" (UID: \"0ef3c015-11a0-43bf-b3db-a4ee755ab04b\") " pod="openshift-dns/dns-default-4kpkf" Apr 22 14:19:29.522456 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.522390 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daaf870a-47fd-4fba-8f87-375c235c1a4c-cert\") pod \"ingress-canary-b6brs\" (UID: \"daaf870a-47fd-4fba-8f87-375c235c1a4c\") " pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:19:29.539749 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.539724 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-qh95w\"" Apr 22 14:19:29.540268 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.540253 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-hhqr7\"" Apr 22 14:19:29.547796 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.547774 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4kpkf" Apr 22 14:19:29.547865 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.547781 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b6brs" Apr 22 14:19:29.698076 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.698037 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4kpkf"] Apr 22 14:19:29.701998 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:19:29.701967 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ef3c015_11a0_43bf_b3db_a4ee755ab04b.slice/crio-255cd51b2eed6d247caf045561a111e5b79b1d9eb331a547616bb4aece8672ac WatchSource:0}: Error finding container 255cd51b2eed6d247caf045561a111e5b79b1d9eb331a547616bb4aece8672ac: Status 404 returned error can't find the container with id 255cd51b2eed6d247caf045561a111e5b79b1d9eb331a547616bb4aece8672ac Apr 22 14:19:29.722973 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.722949 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b6brs"] Apr 22 14:19:29.846772 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.846655 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b6brs" event={"ID":"daaf870a-47fd-4fba-8f87-375c235c1a4c","Type":"ContainerStarted","Data":"99f0cdb8636eeb47514f1c5dd097d7fd16625cb84f038ca02d1eaadb2aee88ab"} Apr 22 14:19:29.847643 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:29.847618 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4kpkf" event={"ID":"0ef3c015-11a0-43bf-b3db-a4ee755ab04b","Type":"ContainerStarted","Data":"255cd51b2eed6d247caf045561a111e5b79b1d9eb331a547616bb4aece8672ac"} Apr 22 14:19:32.859392 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:32.859350 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b6brs" event={"ID":"daaf870a-47fd-4fba-8f87-375c235c1a4c","Type":"ContainerStarted","Data":"aa2051e43478ba992e3cfed6cb9c7edc8d3c703076c89280ea21be8b1a0e7c93"} Apr 22 14:19:32.860947 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:32.860925 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4kpkf" event={"ID":"0ef3c015-11a0-43bf-b3db-a4ee755ab04b","Type":"ContainerStarted","Data":"396095aa3e9f503d2127fd44c658e6c91aeb80036a9f277b2328263ec5eec4c4"} Apr 22 14:19:32.861049 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:32.860953 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4kpkf" event={"ID":"0ef3c015-11a0-43bf-b3db-a4ee755ab04b","Type":"ContainerStarted","Data":"33250569d027954dd0d9baa35b04e3d49acd652a89170ea2f00455c04a96799d"} Apr 22 14:19:32.861049 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:32.861044 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-4kpkf" Apr 22 14:19:32.875263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:32.875219 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b6brs" podStartSLOduration=251.706566704 podStartE2EDuration="4m13.875204929s" podCreationTimestamp="2026-04-22 14:15:19 +0000 UTC" firstStartedPulling="2026-04-22 14:19:29.73133437 +0000 UTC m=+283.199212081" lastFinishedPulling="2026-04-22 14:19:31.899972581 +0000 UTC m=+285.367850306" observedRunningTime="2026-04-22 14:19:32.874031851 +0000 UTC m=+286.341909582" watchObservedRunningTime="2026-04-22 14:19:32.875204929 +0000 UTC m=+286.343082659" Apr 22 14:19:32.891886 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:32.891825 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-4kpkf" podStartSLOduration=251.693634147 podStartE2EDuration="4m13.8918078s" podCreationTimestamp="2026-04-22 14:15:19 +0000 UTC" firstStartedPulling="2026-04-22 14:19:29.703757519 +0000 UTC m=+283.171635230" lastFinishedPulling="2026-04-22 14:19:31.901931171 +0000 UTC m=+285.369808883" observedRunningTime="2026-04-22 14:19:32.89168451 +0000 UTC m=+286.359562272" watchObservedRunningTime="2026-04-22 14:19:32.8918078 +0000 UTC m=+286.359685531" Apr 22 14:19:42.866252 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:42.866223 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-4kpkf" Apr 22 14:19:46.904889 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:46.904856 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:19:46.905386 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:46.905000 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:19:46.910408 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:46.910385 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:19:46.910548 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:46.910389 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:19:46.918496 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:19:46.918475 2574 kubelet.go:1628] "Image garbage collection succeeded" Apr 22 14:22:21.355030 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.354933 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-c92rf"] Apr 22 14:22:21.355477 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.355268 2574 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" containerName="console" Apr 22 14:22:21.355477 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.355278 2574 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" containerName="console" Apr 22 14:22:21.355477 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.355295 2574 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c6863efe-089d-4c10-a59b-7cc11a649cc1" containerName="console" Apr 22 14:22:21.355477 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.355301 2574 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6863efe-089d-4c10-a59b-7cc11a649cc1" containerName="console" Apr 22 14:22:21.355477 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.355354 2574 memory_manager.go:356] "RemoveStaleState removing state" podUID="b79bdabc-7a55-43ee-a856-0cbc5b1c49c6" containerName="console" Apr 22 14:22:21.355477 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.355363 2574 memory_manager.go:356] "RemoveStaleState removing state" podUID="c6863efe-089d-4c10-a59b-7cc11a649cc1" containerName="console" Apr 22 14:22:21.358371 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.358351 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.361793 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.361769 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Apr 22 14:22:21.361946 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.361769 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Apr 22 14:22:21.361946 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.361773 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-5xbwv\"" Apr 22 14:22:21.369957 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.369936 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-c92rf"] Apr 22 14:22:21.398351 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.398322 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7cr8\" (UniqueName: \"kubernetes.io/projected/f542b752-ce8a-4441-94ce-1f16bb5a725c-kube-api-access-r7cr8\") pod \"cert-manager-webhook-587ccfb98-c92rf\" (UID: \"f542b752-ce8a-4441-94ce-1f16bb5a725c\") " pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.398482 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.398384 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f542b752-ce8a-4441-94ce-1f16bb5a725c-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-c92rf\" (UID: \"f542b752-ce8a-4441-94ce-1f16bb5a725c\") " pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.498893 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.498853 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f542b752-ce8a-4441-94ce-1f16bb5a725c-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-c92rf\" (UID: \"f542b752-ce8a-4441-94ce-1f16bb5a725c\") " pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.499075 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.498960 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r7cr8\" (UniqueName: \"kubernetes.io/projected/f542b752-ce8a-4441-94ce-1f16bb5a725c-kube-api-access-r7cr8\") pod \"cert-manager-webhook-587ccfb98-c92rf\" (UID: \"f542b752-ce8a-4441-94ce-1f16bb5a725c\") " pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.510163 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.510129 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f542b752-ce8a-4441-94ce-1f16bb5a725c-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-c92rf\" (UID: \"f542b752-ce8a-4441-94ce-1f16bb5a725c\") " pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.510291 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.510192 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7cr8\" (UniqueName: \"kubernetes.io/projected/f542b752-ce8a-4441-94ce-1f16bb5a725c-kube-api-access-r7cr8\") pod \"cert-manager-webhook-587ccfb98-c92rf\" (UID: \"f542b752-ce8a-4441-94ce-1f16bb5a725c\") " pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.676811 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.676779 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:21.815512 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.815417 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-c92rf"] Apr 22 14:22:21.818217 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:22:21.818188 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf542b752_ce8a_4441_94ce_1f16bb5a725c.slice/crio-2284441c7875a11e28e8e682791301bef04e594898d24279e5a2c9e377665e75 WatchSource:0}: Error finding container 2284441c7875a11e28e8e682791301bef04e594898d24279e5a2c9e377665e75: Status 404 returned error can't find the container with id 2284441c7875a11e28e8e682791301bef04e594898d24279e5a2c9e377665e75 Apr 22 14:22:21.820104 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:21.820089 2574 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 14:22:22.365107 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:22.365057 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" event={"ID":"f542b752-ce8a-4441-94ce-1f16bb5a725c","Type":"ContainerStarted","Data":"2284441c7875a11e28e8e682791301bef04e594898d24279e5a2c9e377665e75"} Apr 22 14:22:24.206033 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.205990 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-68b757865b-kd82h"] Apr 22 14:22:24.209867 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.209826 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:24.212189 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.212168 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-ls8xx\"" Apr 22 14:22:24.219321 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.219193 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-68b757865b-kd82h"] Apr 22 14:22:24.224801 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.224775 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r9xr\" (UniqueName: \"kubernetes.io/projected/4b2f69ce-b4bd-4840-b4d4-8e878823ba0c-kube-api-access-9r9xr\") pod \"cert-manager-cainjector-68b757865b-kd82h\" (UID: \"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:24.224921 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.224830 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b2f69ce-b4bd-4840-b4d4-8e878823ba0c-bound-sa-token\") pod \"cert-manager-cainjector-68b757865b-kd82h\" (UID: \"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:24.326196 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.326159 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9r9xr\" (UniqueName: \"kubernetes.io/projected/4b2f69ce-b4bd-4840-b4d4-8e878823ba0c-kube-api-access-9r9xr\") pod \"cert-manager-cainjector-68b757865b-kd82h\" (UID: \"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:24.326379 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.326231 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b2f69ce-b4bd-4840-b4d4-8e878823ba0c-bound-sa-token\") pod \"cert-manager-cainjector-68b757865b-kd82h\" (UID: \"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:24.335949 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.335920 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b2f69ce-b4bd-4840-b4d4-8e878823ba0c-bound-sa-token\") pod \"cert-manager-cainjector-68b757865b-kd82h\" (UID: \"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:24.336099 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.336003 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r9xr\" (UniqueName: \"kubernetes.io/projected/4b2f69ce-b4bd-4840-b4d4-8e878823ba0c-kube-api-access-9r9xr\") pod \"cert-manager-cainjector-68b757865b-kd82h\" (UID: \"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:24.522447 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:24.522343 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" Apr 22 14:22:25.037481 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:25.037447 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-68b757865b-kd82h"] Apr 22 14:22:25.040317 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:22:25.040296 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b2f69ce_b4bd_4840_b4d4_8e878823ba0c.slice/crio-24a88cb8f1b293966d364760e78e5c38561aeb640160b35cf01747acbfb105a0 WatchSource:0}: Error finding container 24a88cb8f1b293966d364760e78e5c38561aeb640160b35cf01747acbfb105a0: Status 404 returned error can't find the container with id 24a88cb8f1b293966d364760e78e5c38561aeb640160b35cf01747acbfb105a0 Apr 22 14:22:25.377248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:25.377149 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" event={"ID":"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c","Type":"ContainerStarted","Data":"afdeb6e687614f30d37d7a6121aa848ec42e2b13ef214fce62b4c6b3efd8639a"} Apr 22 14:22:25.377248 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:25.377193 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" event={"ID":"4b2f69ce-b4bd-4840-b4d4-8e878823ba0c","Type":"ContainerStarted","Data":"24a88cb8f1b293966d364760e78e5c38561aeb640160b35cf01747acbfb105a0"} Apr 22 14:22:25.378565 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:25.378545 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" event={"ID":"f542b752-ce8a-4441-94ce-1f16bb5a725c","Type":"ContainerStarted","Data":"3593668fb839d755df586e8158981928408cbebc59d2cd405b439c571b4fa530"} Apr 22 14:22:25.378631 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:25.378583 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:22:25.394843 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:25.394792 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-68b757865b-kd82h" podStartSLOduration=1.394778218 podStartE2EDuration="1.394778218s" podCreationTimestamp="2026-04-22 14:22:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:22:25.39228474 +0000 UTC m=+458.860162469" watchObservedRunningTime="2026-04-22 14:22:25.394778218 +0000 UTC m=+458.862655948" Apr 22 14:22:25.411982 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:25.411941 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" podStartSLOduration=1.2578962439999999 podStartE2EDuration="4.411926027s" podCreationTimestamp="2026-04-22 14:22:21 +0000 UTC" firstStartedPulling="2026-04-22 14:22:21.820215313 +0000 UTC m=+455.288093021" lastFinishedPulling="2026-04-22 14:22:24.974245078 +0000 UTC m=+458.442122804" observedRunningTime="2026-04-22 14:22:25.410441252 +0000 UTC m=+458.878318984" watchObservedRunningTime="2026-04-22 14:22:25.411926027 +0000 UTC m=+458.879803760" Apr 22 14:22:31.384720 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:22:31.384656 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-587ccfb98-c92rf" Apr 22 14:23:07.930801 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.930758 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-lws-operator/lws-controller-manager-879f8864c-95qvc"] Apr 22 14:23:07.934287 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.934258 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:07.940758 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.940728 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"kube-root-ca.crt\"" Apr 22 14:23:07.941648 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.941623 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"lws-manager-config\"" Apr 22 14:23:07.942544 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.942525 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"openshift-service-ca.crt\"" Apr 22 14:23:07.942617 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.942551 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"metrics-server-cert\"" Apr 22 14:23:07.942617 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.942534 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"lws-controller-manager-dockercfg-kx596\"" Apr 22 14:23:07.945902 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.945870 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"webhook-server-cert\"" Apr 22 14:23:07.970608 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:07.970565 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-879f8864c-95qvc"] Apr 22 14:23:08.006323 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.006291 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/10673011-5fdc-4c8e-aa42-ce89381bc7f6-manager-config\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.006501 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.006347 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10673011-5fdc-4c8e-aa42-ce89381bc7f6-cert\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.006501 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.006371 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/10673011-5fdc-4c8e-aa42-ce89381bc7f6-metrics-cert\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.006501 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.006398 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqzqt\" (UniqueName: \"kubernetes.io/projected/10673011-5fdc-4c8e-aa42-ce89381bc7f6-kube-api-access-fqzqt\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.107953 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.107906 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10673011-5fdc-4c8e-aa42-ce89381bc7f6-cert\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.107953 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.107967 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/10673011-5fdc-4c8e-aa42-ce89381bc7f6-metrics-cert\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.108228 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.108008 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fqzqt\" (UniqueName: \"kubernetes.io/projected/10673011-5fdc-4c8e-aa42-ce89381bc7f6-kube-api-access-fqzqt\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.108228 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.108123 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/10673011-5fdc-4c8e-aa42-ce89381bc7f6-manager-config\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.108745 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.108720 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/10673011-5fdc-4c8e-aa42-ce89381bc7f6-manager-config\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.111786 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.111758 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/10673011-5fdc-4c8e-aa42-ce89381bc7f6-metrics-cert\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.112329 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.112282 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10673011-5fdc-4c8e-aa42-ce89381bc7f6-cert\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.125250 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.125213 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqzqt\" (UniqueName: \"kubernetes.io/projected/10673011-5fdc-4c8e-aa42-ce89381bc7f6-kube-api-access-fqzqt\") pod \"lws-controller-manager-879f8864c-95qvc\" (UID: \"10673011-5fdc-4c8e-aa42-ce89381bc7f6\") " pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.244038 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.243930 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:08.422183 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.422063 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-879f8864c-95qvc"] Apr 22 14:23:08.425041 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:23:08.425009 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10673011_5fdc_4c8e_aa42_ce89381bc7f6.slice/crio-500d2d79d0fc6cb3430e3fa09828ffdbf1f73d66fcc3f5c5c009702cf0bf0f18 WatchSource:0}: Error finding container 500d2d79d0fc6cb3430e3fa09828ffdbf1f73d66fcc3f5c5c009702cf0bf0f18: Status 404 returned error can't find the container with id 500d2d79d0fc6cb3430e3fa09828ffdbf1f73d66fcc3f5c5c009702cf0bf0f18 Apr 22 14:23:08.513391 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:08.512553 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" event={"ID":"10673011-5fdc-4c8e-aa42-ce89381bc7f6","Type":"ContainerStarted","Data":"500d2d79d0fc6cb3430e3fa09828ffdbf1f73d66fcc3f5c5c009702cf0bf0f18"} Apr 22 14:23:11.522907 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:11.522868 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" event={"ID":"10673011-5fdc-4c8e-aa42-ce89381bc7f6","Type":"ContainerStarted","Data":"339649aade903405e431bfb06f951ddf1cbdc6211639880fc34a8f89953925d9"} Apr 22 14:23:11.523324 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:11.522983 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:11.582640 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:11.582582 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" podStartSLOduration=2.030011187 podStartE2EDuration="4.582565625s" podCreationTimestamp="2026-04-22 14:23:07 +0000 UTC" firstStartedPulling="2026-04-22 14:23:08.426979838 +0000 UTC m=+501.894857547" lastFinishedPulling="2026-04-22 14:23:10.979534277 +0000 UTC m=+504.447411985" observedRunningTime="2026-04-22 14:23:11.580726758 +0000 UTC m=+505.048604488" watchObservedRunningTime="2026-04-22 14:23:11.582565625 +0000 UTC m=+505.050443355" Apr 22 14:23:19.283104 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.283065 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-6fdc578559-nw57p"] Apr 22 14:23:19.286833 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.286805 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.289288 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.289255 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 22 14:23:19.289573 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.289556 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 22 14:23:19.289888 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.289867 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 22 14:23:19.290021 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.289915 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 22 14:23:19.290021 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.289948 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 22 14:23:19.290249 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.290220 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 22 14:23:19.290356 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.290297 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-78jk5\"" Apr 22 14:23:19.290661 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.290645 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 22 14:23:19.296209 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.296188 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 22 14:23:19.303740 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.303711 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6fdc578559-nw57p"] Apr 22 14:23:19.409441 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.409381 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmthn\" (UniqueName: \"kubernetes.io/projected/53f7749b-5440-4127-abd4-17e2936cc3b8-kube-api-access-nmthn\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.409441 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.409453 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-oauth-serving-cert\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.409721 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.409495 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/53f7749b-5440-4127-abd4-17e2936cc3b8-console-serving-cert\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.409721 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.409514 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-console-config\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.409721 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.409647 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/53f7749b-5440-4127-abd4-17e2936cc3b8-console-oauth-config\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.409721 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.409714 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-service-ca\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.409894 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.409740 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-trusted-ca-bundle\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511022 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.510963 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/53f7749b-5440-4127-abd4-17e2936cc3b8-console-serving-cert\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511022 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.511030 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-console-config\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511286 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.511118 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/53f7749b-5440-4127-abd4-17e2936cc3b8-console-oauth-config\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511286 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.511140 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-service-ca\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511286 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.511167 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-trusted-ca-bundle\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511286 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.511193 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nmthn\" (UniqueName: \"kubernetes.io/projected/53f7749b-5440-4127-abd4-17e2936cc3b8-kube-api-access-nmthn\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511286 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.511229 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-oauth-serving-cert\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.511941 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.511917 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-console-config\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.512088 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.512064 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-service-ca\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.512131 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.512103 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-trusted-ca-bundle\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.512272 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.512249 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/53f7749b-5440-4127-abd4-17e2936cc3b8-oauth-serving-cert\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.513669 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.513641 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/53f7749b-5440-4127-abd4-17e2936cc3b8-console-oauth-config\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.513770 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.513695 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/53f7749b-5440-4127-abd4-17e2936cc3b8-console-serving-cert\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.519505 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.519487 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmthn\" (UniqueName: \"kubernetes.io/projected/53f7749b-5440-4127-abd4-17e2936cc3b8-kube-api-access-nmthn\") pod \"console-6fdc578559-nw57p\" (UID: \"53f7749b-5440-4127-abd4-17e2936cc3b8\") " pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.599202 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.599115 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:19.733238 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:19.733210 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6fdc578559-nw57p"] Apr 22 14:23:19.736069 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:23:19.736033 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53f7749b_5440_4127_abd4_17e2936cc3b8.slice/crio-c832452332c557a7672afe63d2f6dbdbfa97b610da50c5a8ef3aeaf7bb653b70 WatchSource:0}: Error finding container c832452332c557a7672afe63d2f6dbdbfa97b610da50c5a8ef3aeaf7bb653b70: Status 404 returned error can't find the container with id c832452332c557a7672afe63d2f6dbdbfa97b610da50c5a8ef3aeaf7bb653b70 Apr 22 14:23:20.552098 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:20.552061 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6fdc578559-nw57p" event={"ID":"53f7749b-5440-4127-abd4-17e2936cc3b8","Type":"ContainerStarted","Data":"ad71951852534e940857265cca6cdb6211aab61b469a90a0c85e17c89db52c73"} Apr 22 14:23:20.552098 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:20.552099 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6fdc578559-nw57p" event={"ID":"53f7749b-5440-4127-abd4-17e2936cc3b8","Type":"ContainerStarted","Data":"c832452332c557a7672afe63d2f6dbdbfa97b610da50c5a8ef3aeaf7bb653b70"} Apr 22 14:23:22.528810 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:22.528778 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-lws-operator/lws-controller-manager-879f8864c-95qvc" Apr 22 14:23:22.546481 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:22.546432 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6fdc578559-nw57p" podStartSLOduration=3.546415663 podStartE2EDuration="3.546415663s" podCreationTimestamp="2026-04-22 14:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:23:20.574400027 +0000 UTC m=+514.042277759" watchObservedRunningTime="2026-04-22 14:23:22.546415663 +0000 UTC m=+516.014293394" Apr 22 14:23:29.600173 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:29.600140 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:29.600173 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:29.600177 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:29.604943 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:29.604915 2574 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:30.589082 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:30.589053 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6fdc578559-nw57p" Apr 22 14:23:47.839706 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.839652 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt"] Apr 22 14:23:47.843347 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.843327 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" Apr 22 14:23:47.845593 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.845576 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"dns-operator-controller-env\"" Apr 22 14:23:47.845745 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.845726 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"dns-operator-controller-manager-dockercfg-kr7nt\"" Apr 22 14:23:47.845985 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.845970 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"kube-root-ca.crt\"" Apr 22 14:23:47.846121 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.846102 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"openshift-service-ca.crt\"" Apr 22 14:23:47.860104 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.860080 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt"] Apr 22 14:23:47.972515 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:47.972477 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74zhb\" (UniqueName: \"kubernetes.io/projected/90e6deaa-aa78-4296-82c6-c06e2200a4a1-kube-api-access-74zhb\") pod \"dns-operator-controller-manager-844548ff4c-qf7lt\" (UID: \"90e6deaa-aa78-4296-82c6-c06e2200a4a1\") " pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" Apr 22 14:23:48.074001 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:48.073957 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-74zhb\" (UniqueName: \"kubernetes.io/projected/90e6deaa-aa78-4296-82c6-c06e2200a4a1-kube-api-access-74zhb\") pod \"dns-operator-controller-manager-844548ff4c-qf7lt\" (UID: \"90e6deaa-aa78-4296-82c6-c06e2200a4a1\") " pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" Apr 22 14:23:48.083217 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:48.083185 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-74zhb\" (UniqueName: \"kubernetes.io/projected/90e6deaa-aa78-4296-82c6-c06e2200a4a1-kube-api-access-74zhb\") pod \"dns-operator-controller-manager-844548ff4c-qf7lt\" (UID: \"90e6deaa-aa78-4296-82c6-c06e2200a4a1\") " pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" Apr 22 14:23:48.153699 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:48.153647 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" Apr 22 14:23:48.314754 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:48.314722 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt"] Apr 22 14:23:48.318242 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:23:48.318209 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90e6deaa_aa78_4296_82c6_c06e2200a4a1.slice/crio-d4f47dc502a0a90419111fc393f385e4878508fc2ff87830217863397848b4d9 WatchSource:0}: Error finding container d4f47dc502a0a90419111fc393f385e4878508fc2ff87830217863397848b4d9: Status 404 returned error can't find the container with id d4f47dc502a0a90419111fc393f385e4878508fc2ff87830217863397848b4d9 Apr 22 14:23:48.644309 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:48.644277 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" event={"ID":"90e6deaa-aa78-4296-82c6-c06e2200a4a1","Type":"ContainerStarted","Data":"d4f47dc502a0a90419111fc393f385e4878508fc2ff87830217863397848b4d9"} Apr 22 14:23:51.660000 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:51.659949 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" event={"ID":"90e6deaa-aa78-4296-82c6-c06e2200a4a1","Type":"ContainerStarted","Data":"ac8421a9e8a213eb8b7a3964d93e90132794c50a12bbfdee4b177505f44359d8"} Apr 22 14:23:51.660428 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:51.660038 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" Apr 22 14:23:51.695067 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:51.695014 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" podStartSLOduration=2.199370972 podStartE2EDuration="4.695000661s" podCreationTimestamp="2026-04-22 14:23:47 +0000 UTC" firstStartedPulling="2026-04-22 14:23:48.32031055 +0000 UTC m=+541.788188258" lastFinishedPulling="2026-04-22 14:23:50.815940236 +0000 UTC m=+544.283817947" observedRunningTime="2026-04-22 14:23:51.691964234 +0000 UTC m=+545.159841965" watchObservedRunningTime="2026-04-22 14:23:51.695000661 +0000 UTC m=+545.162878388" Apr 22 14:23:52.858379 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:52.858345 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml"] Apr 22 14:23:52.861813 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:52.861783 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" Apr 22 14:23:52.864870 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:52.864826 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"limitador-operator-controller-manager-dockercfg-cnrk7\"" Apr 22 14:23:52.879282 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:52.879247 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml"] Apr 22 14:23:52.920835 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:52.920803 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qc5c\" (UniqueName: \"kubernetes.io/projected/9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38-kube-api-access-6qc5c\") pod \"limitador-operator-controller-manager-c7fb4c8d5-l56ml\" (UID: \"9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38\") " pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" Apr 22 14:23:53.022056 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:53.022022 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6qc5c\" (UniqueName: \"kubernetes.io/projected/9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38-kube-api-access-6qc5c\") pod \"limitador-operator-controller-manager-c7fb4c8d5-l56ml\" (UID: \"9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38\") " pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" Apr 22 14:23:53.047550 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:53.047520 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qc5c\" (UniqueName: \"kubernetes.io/projected/9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38-kube-api-access-6qc5c\") pod \"limitador-operator-controller-manager-c7fb4c8d5-l56ml\" (UID: \"9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38\") " pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" Apr 22 14:23:53.172481 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:53.172444 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" Apr 22 14:23:53.299778 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:53.299739 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml"] Apr 22 14:23:53.303215 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:23:53.303172 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9704f09b_ac9b_493f_bf6a_0b5ef0b0fa38.slice/crio-d0c06c8add8508c8a5e83043de45807fd74a6a8afd2dd96ff38354e1f19d700b WatchSource:0}: Error finding container d0c06c8add8508c8a5e83043de45807fd74a6a8afd2dd96ff38354e1f19d700b: Status 404 returned error can't find the container with id d0c06c8add8508c8a5e83043de45807fd74a6a8afd2dd96ff38354e1f19d700b Apr 22 14:23:53.670595 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:53.670561 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" event={"ID":"9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38","Type":"ContainerStarted","Data":"d0c06c8add8508c8a5e83043de45807fd74a6a8afd2dd96ff38354e1f19d700b"} Apr 22 14:23:55.681123 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:55.681087 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" event={"ID":"9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38","Type":"ContainerStarted","Data":"0a413e687d9695168b450b18cd724bb18c164f42c1246ccc038717acad79e3cc"} Apr 22 14:23:55.681611 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:55.681297 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" Apr 22 14:23:55.724134 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:23:55.724072 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" podStartSLOduration=2.07020573 podStartE2EDuration="3.724054974s" podCreationTimestamp="2026-04-22 14:23:52 +0000 UTC" firstStartedPulling="2026-04-22 14:23:53.305702377 +0000 UTC m=+546.773580098" lastFinishedPulling="2026-04-22 14:23:54.959551622 +0000 UTC m=+548.427429342" observedRunningTime="2026-04-22 14:23:55.71774718 +0000 UTC m=+549.185624913" watchObservedRunningTime="2026-04-22 14:23:55.724054974 +0000 UTC m=+549.191932705" Apr 22 14:24:02.669230 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:02.669193 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/dns-operator-controller-manager-844548ff4c-qf7lt" Apr 22 14:24:06.687449 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:06.687419 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/limitador-operator-controller-manager-c7fb4c8d5-l56ml" Apr 22 14:24:45.498400 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.498360 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/limitador-limitador-64c8f475fb-hc7j2"] Apr 22 14:24:45.500688 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.500657 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.503766 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.503747 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"limitador-limits-config-limitador\"" Apr 22 14:24:45.503766 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.503755 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"default-dockercfg-pkxlf\"" Apr 22 14:24:45.518047 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.518019 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-64c8f475fb-hc7j2"] Apr 22 14:24:45.590341 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.590306 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67d6d\" (UniqueName: \"kubernetes.io/projected/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-kube-api-access-67d6d\") pod \"limitador-limitador-64c8f475fb-hc7j2\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.590500 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.590352 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-config-file\") pod \"limitador-limitador-64c8f475fb-hc7j2\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.597405 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.597377 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-64c8f475fb-hc7j2"] Apr 22 14:24:45.691854 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.691813 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-67d6d\" (UniqueName: \"kubernetes.io/projected/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-kube-api-access-67d6d\") pod \"limitador-limitador-64c8f475fb-hc7j2\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.692032 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.691861 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-config-file\") pod \"limitador-limitador-64c8f475fb-hc7j2\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.692571 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.692552 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-config-file\") pod \"limitador-limitador-64c8f475fb-hc7j2\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.701247 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.701219 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-67d6d\" (UniqueName: \"kubernetes.io/projected/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-kube-api-access-67d6d\") pod \"limitador-limitador-64c8f475fb-hc7j2\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.811505 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.811420 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:45.943185 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:45.943158 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-64c8f475fb-hc7j2"] Apr 22 14:24:45.945362 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:24:45.945330 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf12ee9df_81ba_4a57_a3c5_8f9ed626bc6a.slice/crio-fdb618c8113dae9258931ba00f06d4bcef2ab76899e0b4adc7544859839a35ff WatchSource:0}: Error finding container fdb618c8113dae9258931ba00f06d4bcef2ab76899e0b4adc7544859839a35ff: Status 404 returned error can't find the container with id fdb618c8113dae9258931ba00f06d4bcef2ab76899e0b4adc7544859839a35ff Apr 22 14:24:46.868339 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:46.868297 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" event={"ID":"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a","Type":"ContainerStarted","Data":"fdb618c8113dae9258931ba00f06d4bcef2ab76899e0b4adc7544859839a35ff"} Apr 22 14:24:46.936324 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:46.936288 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:24:46.939150 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:46.939126 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:24:46.941259 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:46.941236 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:24:46.946508 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:46.946481 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:24:49.881096 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:49.881060 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" event={"ID":"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a","Type":"ContainerStarted","Data":"b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7"} Apr 22 14:24:49.881469 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:49.881225 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:24:49.904301 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:24:49.904243 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" podStartSLOduration=1.133414595 podStartE2EDuration="4.904225459s" podCreationTimestamp="2026-04-22 14:24:45 +0000 UTC" firstStartedPulling="2026-04-22 14:24:45.947383467 +0000 UTC m=+599.415261192" lastFinishedPulling="2026-04-22 14:24:49.718194334 +0000 UTC m=+603.186072056" observedRunningTime="2026-04-22 14:24:49.902585652 +0000 UTC m=+603.370463382" watchObservedRunningTime="2026-04-22 14:24:49.904225459 +0000 UTC m=+603.372103192" Apr 22 14:25:00.885585 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:00.885554 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:25:01.831811 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:01.831775 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/limitador-limitador-64c8f475fb-hc7j2"] Apr 22 14:25:01.832009 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:01.831985 2574 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" podUID="f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" containerName="limitador" containerID="cri-o://b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7" gracePeriod=30 Apr 22 14:25:02.379932 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.379907 2574 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:25:02.444717 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.444607 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-config-file\") pod \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " Apr 22 14:25:02.444888 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.444716 2574 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67d6d\" (UniqueName: \"kubernetes.io/projected/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-kube-api-access-67d6d\") pod \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\" (UID: \"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a\") " Apr 22 14:25:02.444974 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.444945 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-config-file" (OuterVolumeSpecName: "config-file") pod "f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" (UID: "f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a"). InnerVolumeSpecName "config-file". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 14:25:02.447026 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.447004 2574 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-kube-api-access-67d6d" (OuterVolumeSpecName: "kube-api-access-67d6d") pod "f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" (UID: "f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a"). InnerVolumeSpecName "kube-api-access-67d6d". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 14:25:02.546096 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.546059 2574 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-67d6d\" (UniqueName: \"kubernetes.io/projected/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-kube-api-access-67d6d\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:25:02.546096 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.546090 2574 reconciler_common.go:299] "Volume detached for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a-config-file\") on node \"ip-10-0-136-18.ec2.internal\" DevicePath \"\"" Apr 22 14:25:02.928437 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.928402 2574 generic.go:358] "Generic (PLEG): container finished" podID="f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" containerID="b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7" exitCode=0 Apr 22 14:25:02.928605 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.928470 2574 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" Apr 22 14:25:02.928605 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.928485 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" event={"ID":"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a","Type":"ContainerDied","Data":"b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7"} Apr 22 14:25:02.928605 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.928523 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-64c8f475fb-hc7j2" event={"ID":"f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a","Type":"ContainerDied","Data":"fdb618c8113dae9258931ba00f06d4bcef2ab76899e0b4adc7544859839a35ff"} Apr 22 14:25:02.928605 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.928541 2574 scope.go:117] "RemoveContainer" containerID="b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7" Apr 22 14:25:02.937170 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.937150 2574 scope.go:117] "RemoveContainer" containerID="b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7" Apr 22 14:25:02.937409 ip-10-0-136-18 kubenswrapper[2574]: E0422 14:25:02.937392 2574 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7\": container with ID starting with b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7 not found: ID does not exist" containerID="b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7" Apr 22 14:25:02.937457 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.937418 2574 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7"} err="failed to get container status \"b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7\": rpc error: code = NotFound desc = could not find container \"b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7\": container with ID starting with b266ff12f08592ad0063ef8975927ff7e8ed6f1261cb17ca1d8898bc465e26f7 not found: ID does not exist" Apr 22 14:25:02.965058 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.965017 2574 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/limitador-limitador-64c8f475fb-hc7j2"] Apr 22 14:25:02.968242 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.968217 2574 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/limitador-limitador-64c8f475fb-hc7j2"] Apr 22 14:25:02.993482 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:02.993442 2574 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" path="/var/lib/kubelet/pods/f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a/volumes" Apr 22 14:25:21.775448 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.775411 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv"] Apr 22 14:25:21.775971 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.775951 2574 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" containerName="limitador" Apr 22 14:25:21.776046 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.775975 2574 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" containerName="limitador" Apr 22 14:25:21.776098 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.776081 2574 memory_manager.go:356] "RemoveStaleState removing state" podUID="f12ee9df-81ba-4a57-a3c5-8f9ed626bc6a" containerName="limitador" Apr 22 14:25:21.789600 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.789566 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:21.794915 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.794886 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"cacerts\"" Apr 22 14:25:21.795256 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.795231 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"istiod-tls\"" Apr 22 14:25:21.795355 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.795279 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"istio-kubeconfig\"" Apr 22 14:25:21.795355 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.795297 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Apr 22 14:25:21.795471 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.795297 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"istio-ca-root-cert\"" Apr 22 14:25:21.795726 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.795574 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"istiod-openshift-gateway-dockercfg-rkms6\"" Apr 22 14:25:21.795726 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.795598 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Apr 22 14:25:21.796427 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.796408 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv"] Apr 22 14:25:21.928067 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.928025 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bnth\" (UniqueName: \"kubernetes.io/projected/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-kube-api-access-7bnth\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:21.928241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.928073 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-kubeconfig\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-kubeconfig\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:21.928241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.928142 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-certs\" (UniqueName: \"kubernetes.io/empty-dir/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-local-certs\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:21.928241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.928181 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-token\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:21.928241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.928198 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cacerts\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-cacerts\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:21.928370 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.928247 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-csr-ca-configmap\" (UniqueName: \"kubernetes.io/configmap/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-csr-ca-configmap\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:21.928370 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:21.928285 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-csr-dns-cert\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-csr-dns-cert\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.029012 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.028919 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7bnth\" (UniqueName: \"kubernetes.io/projected/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-kube-api-access-7bnth\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.029012 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.028959 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-kubeconfig\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-kubeconfig\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.029012 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.028988 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"local-certs\" (UniqueName: \"kubernetes.io/empty-dir/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-local-certs\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.029012 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.029013 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-token\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.029304 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.029141 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cacerts\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-cacerts\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.029304 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.029226 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-csr-ca-configmap\" (UniqueName: \"kubernetes.io/configmap/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-csr-ca-configmap\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.029304 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.029275 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-csr-dns-cert\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-csr-dns-cert\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.030040 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.030004 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-csr-ca-configmap\" (UniqueName: \"kubernetes.io/configmap/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-csr-ca-configmap\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.031615 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.031587 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"local-certs\" (UniqueName: \"kubernetes.io/empty-dir/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-local-certs\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.031744 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.031636 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-csr-dns-cert\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-csr-dns-cert\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.031910 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.031893 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cacerts\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-cacerts\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.032003 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.031969 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-kubeconfig\" (UniqueName: \"kubernetes.io/secret/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-kubeconfig\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.043256 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.043233 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-istio-token\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.056858 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.056827 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bnth\" (UniqueName: \"kubernetes.io/projected/a2ec9ada-974f-4fa8-80fe-f7a1798b0fff-kube-api-access-7bnth\") pod \"istiod-openshift-gateway-55ff986f96-7cccv\" (UID: \"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff\") " pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.099713 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.099659 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:22.257320 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.252697 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv"] Apr 22 14:25:22.261734 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:25:22.261699 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2ec9ada_974f_4fa8_80fe_f7a1798b0fff.slice/crio-a1a2d548d058eb28c6b5789b8029e31c61db9a2acc7d526442e92a5377082fd4 WatchSource:0}: Error finding container a1a2d548d058eb28c6b5789b8029e31c61db9a2acc7d526442e92a5377082fd4: Status 404 returned error can't find the container with id a1a2d548d058eb28c6b5789b8029e31c61db9a2acc7d526442e92a5377082fd4 Apr 22 14:25:22.997194 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:22.997154 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" event={"ID":"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff","Type":"ContainerStarted","Data":"a1a2d548d058eb28c6b5789b8029e31c61db9a2acc7d526442e92a5377082fd4"} Apr 22 14:25:24.929209 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:24.928658 2574 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892160Ki","pods":"250"} Apr 22 14:25:24.929209 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:24.928751 2574 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"30892160Ki","pods":"250"} Apr 22 14:25:25.006869 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:25.006832 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" event={"ID":"a2ec9ada-974f-4fa8-80fe-f7a1798b0fff","Type":"ContainerStarted","Data":"47b4b5d3b66afd3bca0a03c2a645afffb259938653f153a9aeb90b892872cecc"} Apr 22 14:25:25.007010 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:25.006990 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:25.008514 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:25.008490 2574 patch_prober.go:28] interesting pod/istiod-openshift-gateway-55ff986f96-7cccv container/discovery namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://10.132.0.32:8080/ready\": dial tcp 10.132.0.32:8080: connect: connection refused" start-of-body= Apr 22 14:25:25.008610 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:25.008540 2574 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" podUID="a2ec9ada-974f-4fa8-80fe-f7a1798b0fff" containerName="discovery" probeResult="failure" output="Get \"http://10.132.0.32:8080/ready\": dial tcp 10.132.0.32:8080: connect: connection refused" Apr 22 14:25:26.012617 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:26.012589 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" Apr 22 14:25:26.038344 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:25:26.038283 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/istiod-openshift-gateway-55ff986f96-7cccv" podStartSLOduration=2.374108925 podStartE2EDuration="5.03826296s" podCreationTimestamp="2026-04-22 14:25:21 +0000 UTC" firstStartedPulling="2026-04-22 14:25:22.26418581 +0000 UTC m=+635.732063526" lastFinishedPulling="2026-04-22 14:25:24.928339845 +0000 UTC m=+638.396217561" observedRunningTime="2026-04-22 14:25:25.036765494 +0000 UTC m=+638.504643224" watchObservedRunningTime="2026-04-22 14:25:26.03826296 +0000 UTC m=+639.506140690" Apr 22 14:29:46.976123 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:29:46.975896 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:29:46.979072 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:29:46.979053 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:29:46.980337 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:29:46.980315 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:29:46.982842 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:29:46.982823 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:34:47.002606 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:47.002578 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:34:47.008602 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:47.008575 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:34:47.009247 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:47.009227 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:34:47.012539 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:47.012515 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:34:59.963610 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:59.963571 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve/llmisvc-controller-manager-6649659bc5-p7vvj"] Apr 22 14:34:59.965966 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:59.965945 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:34:59.968613 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:59.968594 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"llmisvc-controller-manager-dockercfg-2lkr9\"" Apr 22 14:34:59.968613 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:59.968605 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"llmisvc-webhook-server-cert\"" Apr 22 14:34:59.968791 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:59.968610 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kserve\"/\"kube-root-ca.crt\"" Apr 22 14:34:59.968791 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:59.968655 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kserve\"/\"openshift-service-ca.crt\"" Apr 22 14:34:59.978506 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:34:59.978486 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/llmisvc-controller-manager-6649659bc5-p7vvj"] Apr 22 14:35:00.118499 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.118464 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8cv4\" (UniqueName: \"kubernetes.io/projected/9fd883d2-6b84-4233-9b58-e0bda80ad8cc-kube-api-access-f8cv4\") pod \"llmisvc-controller-manager-6649659bc5-p7vvj\" (UID: \"9fd883d2-6b84-4233-9b58-e0bda80ad8cc\") " pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:00.118667 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.118632 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9fd883d2-6b84-4233-9b58-e0bda80ad8cc-cert\") pod \"llmisvc-controller-manager-6649659bc5-p7vvj\" (UID: \"9fd883d2-6b84-4233-9b58-e0bda80ad8cc\") " pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:00.220134 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.220059 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9fd883d2-6b84-4233-9b58-e0bda80ad8cc-cert\") pod \"llmisvc-controller-manager-6649659bc5-p7vvj\" (UID: \"9fd883d2-6b84-4233-9b58-e0bda80ad8cc\") " pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:00.220134 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.220102 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f8cv4\" (UniqueName: \"kubernetes.io/projected/9fd883d2-6b84-4233-9b58-e0bda80ad8cc-kube-api-access-f8cv4\") pod \"llmisvc-controller-manager-6649659bc5-p7vvj\" (UID: \"9fd883d2-6b84-4233-9b58-e0bda80ad8cc\") " pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:00.222405 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.222376 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9fd883d2-6b84-4233-9b58-e0bda80ad8cc-cert\") pod \"llmisvc-controller-manager-6649659bc5-p7vvj\" (UID: \"9fd883d2-6b84-4233-9b58-e0bda80ad8cc\") " pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:00.227777 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.227753 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8cv4\" (UniqueName: \"kubernetes.io/projected/9fd883d2-6b84-4233-9b58-e0bda80ad8cc-kube-api-access-f8cv4\") pod \"llmisvc-controller-manager-6649659bc5-p7vvj\" (UID: \"9fd883d2-6b84-4233-9b58-e0bda80ad8cc\") " pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:00.276018 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.275992 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:00.400029 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.400007 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/llmisvc-controller-manager-6649659bc5-p7vvj"] Apr 22 14:35:00.402021 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:35:00.401991 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9fd883d2_6b84_4233_9b58_e0bda80ad8cc.slice/crio-d2923810db308f781220d7a0ce4425891a87da98a7ec6eefd072b2220abc4c0e WatchSource:0}: Error finding container d2923810db308f781220d7a0ce4425891a87da98a7ec6eefd072b2220abc4c0e: Status 404 returned error can't find the container with id d2923810db308f781220d7a0ce4425891a87da98a7ec6eefd072b2220abc4c0e Apr 22 14:35:00.403445 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.403425 2574 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 14:35:00.997999 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:00.997967 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" event={"ID":"9fd883d2-6b84-4233-9b58-e0bda80ad8cc","Type":"ContainerStarted","Data":"d2923810db308f781220d7a0ce4425891a87da98a7ec6eefd072b2220abc4c0e"} Apr 22 14:35:04.010114 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:04.010075 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" event={"ID":"9fd883d2-6b84-4233-9b58-e0bda80ad8cc","Type":"ContainerStarted","Data":"bcc0e0402321c61001e8f17b2297714555ddb3476923b05978af5f8613403847"} Apr 22 14:35:04.010483 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:04.010149 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:35:04.026239 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:04.026184 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" podStartSLOduration=1.629763317 podStartE2EDuration="5.02616844s" podCreationTimestamp="2026-04-22 14:34:59 +0000 UTC" firstStartedPulling="2026-04-22 14:35:00.403606542 +0000 UTC m=+1213.871484263" lastFinishedPulling="2026-04-22 14:35:03.800011677 +0000 UTC m=+1217.267889386" observedRunningTime="2026-04-22 14:35:04.02478455 +0000 UTC m=+1217.492662292" watchObservedRunningTime="2026-04-22 14:35:04.02616844 +0000 UTC m=+1217.494046208" Apr 22 14:35:35.015254 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:35:35.015186 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve/llmisvc-controller-manager-6649659bc5-p7vvj" Apr 22 14:39:47.030820 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:39:47.030787 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:39:47.034622 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:39:47.034598 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:39:47.035281 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:39:47.035263 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:39:47.038788 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:39:47.038773 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:44:47.056218 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:44:47.056190 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:44:47.063440 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:44:47.063417 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:44:47.064560 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:44:47.064537 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:44:47.068385 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:44:47.068363 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:47:22.106868 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:22.106831 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_istiod-openshift-gateway-55ff986f96-7cccv_a2ec9ada-974f-4fa8-80fe-f7a1798b0fff/discovery/0.log" Apr 22 14:47:22.969941 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:22.969898 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_istiod-openshift-gateway-55ff986f96-7cccv_a2ec9ada-974f-4fa8-80fe-f7a1798b0fff/discovery/0.log" Apr 22 14:47:23.874837 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:23.874807 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_dns-operator-controller-manager-844548ff4c-qf7lt_90e6deaa-aa78-4296-82c6-c06e2200a4a1/manager/0.log" Apr 22 14:47:24.022471 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:24.022437 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_limitador-operator-controller-manager-c7fb4c8d5-l56ml_9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38/manager/0.log" Apr 22 14:47:29.654834 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:29.654797 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-zfm4b_3d7a8b58-1717-408a-9901-09f287c3731b/global-pull-secret-syncer/0.log" Apr 22 14:47:29.690281 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:29.690254 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-f42dv_4cf8ecc7-e6d0-4d13-9d56-8e329378e84a/konnectivity-agent/0.log" Apr 22 14:47:29.852197 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:29.852172 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-136-18.ec2.internal_17be8cc2ca560f8da5581c8d587a8582/haproxy/0.log" Apr 22 14:47:33.988832 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:33.988805 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_dns-operator-controller-manager-844548ff4c-qf7lt_90e6deaa-aa78-4296-82c6-c06e2200a4a1/manager/0.log" Apr 22 14:47:34.140241 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:34.140209 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_limitador-operator-controller-manager-c7fb4c8d5-l56ml_9704f09b-ac9b-493f-bf6a-0b5ef0b0fa38/manager/0.log" Apr 22 14:47:35.051881 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.051849 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_19482d30-2da5-4d8a-89da-515e692e97c7/alertmanager/0.log" Apr 22 14:47:35.075327 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.075297 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_19482d30-2da5-4d8a-89da-515e692e97c7/config-reloader/0.log" Apr 22 14:47:35.101722 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.101686 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_19482d30-2da5-4d8a-89da-515e692e97c7/kube-rbac-proxy-web/0.log" Apr 22 14:47:35.134645 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.134620 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_19482d30-2da5-4d8a-89da-515e692e97c7/kube-rbac-proxy/0.log" Apr 22 14:47:35.164579 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.164556 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_19482d30-2da5-4d8a-89da-515e692e97c7/kube-rbac-proxy-metric/0.log" Apr 22 14:47:35.187644 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.187624 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_19482d30-2da5-4d8a-89da-515e692e97c7/prom-label-proxy/0.log" Apr 22 14:47:35.209517 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.209481 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_19482d30-2da5-4d8a-89da-515e692e97c7/init-config-reloader/0.log" Apr 22 14:47:35.284522 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.284446 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-69db897b98-sjd66_eeeda31a-52ab-4389-b06c-eea0f137d09f/kube-state-metrics/0.log" Apr 22 14:47:35.310226 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.310192 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-69db897b98-sjd66_eeeda31a-52ab-4389-b06c-eea0f137d09f/kube-rbac-proxy-main/0.log" Apr 22 14:47:35.329093 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.329069 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-69db897b98-sjd66_eeeda31a-52ab-4389-b06c-eea0f137d09f/kube-rbac-proxy-self/0.log" Apr 22 14:47:35.352081 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.352057 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_metrics-server-b95c9c57c-c2st2_2e68eca2-c4cf-4ccf-994b-69c0a0b47de4/metrics-server/0.log" Apr 22 14:47:35.547264 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.547198 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-bzjx9_d148eabc-a2af-41d8-ba7d-039b9da0132a/node-exporter/0.log" Apr 22 14:47:35.565690 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.565660 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-bzjx9_d148eabc-a2af-41d8-ba7d-039b9da0132a/kube-rbac-proxy/0.log" Apr 22 14:47:35.584247 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.584225 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-bzjx9_d148eabc-a2af-41d8-ba7d-039b9da0132a/init-textfile/0.log" Apr 22 14:47:35.617570 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.617550 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-lvzv8_82e2b9ba-996b-4f92-b967-0052faa65d73/kube-rbac-proxy-main/0.log" Apr 22 14:47:35.636215 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.636197 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-lvzv8_82e2b9ba-996b-4f92-b967-0052faa65d73/kube-rbac-proxy-self/0.log" Apr 22 14:47:35.656959 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.656941 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-lvzv8_82e2b9ba-996b-4f92-b967-0052faa65d73/openshift-state-metrics/0.log" Apr 22 14:47:35.697149 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.697128 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_114db3c0-b406-4a1d-b927-efb1d4fd0e66/prometheus/0.log" Apr 22 14:47:35.718720 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.718701 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_114db3c0-b406-4a1d-b927-efb1d4fd0e66/config-reloader/0.log" Apr 22 14:47:35.739421 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.739395 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_114db3c0-b406-4a1d-b927-efb1d4fd0e66/thanos-sidecar/0.log" Apr 22 14:47:35.761809 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.761790 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_114db3c0-b406-4a1d-b927-efb1d4fd0e66/kube-rbac-proxy-web/0.log" Apr 22 14:47:35.785871 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.785845 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_114db3c0-b406-4a1d-b927-efb1d4fd0e66/kube-rbac-proxy/0.log" Apr 22 14:47:35.806231 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.806177 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_114db3c0-b406-4a1d-b927-efb1d4fd0e66/kube-rbac-proxy-thanos/0.log" Apr 22 14:47:35.826429 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:35.826398 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_114db3c0-b406-4a1d-b927-efb1d4fd0e66/init-config-reloader/0.log" Apr 22 14:47:37.353697 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:37.353602 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-console_networking-console-plugin-cb95c66f6-8fwqv_6513f255-e88c-4c51-8d5b-17018a5c835d/networking-console-plugin/0.log" Apr 22 14:47:37.881382 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:37.881355 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/1.log" Apr 22 14:47:37.888840 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:37.888812 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-pxmjh_00643d7f-54c7-4ca9-8d93-e7cf0c7554c9/console-operator/2.log" Apr 22 14:47:38.340021 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.339987 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6fdc578559-nw57p_53f7749b-5440-4127-abd4-17e2936cc3b8/console/0.log" Apr 22 14:47:38.751782 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.751746 2574 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz"] Apr 22 14:47:38.754333 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.754313 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:38.756563 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.756541 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-54rh7\"/\"kube-root-ca.crt\"" Apr 22 14:47:38.756638 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.756608 2574 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-54rh7\"/\"openshift-service-ca.crt\"" Apr 22 14:47:38.757267 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.757250 2574 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-54rh7\"/\"default-dockercfg-4zg56\"" Apr 22 14:47:38.768580 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.768558 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz"] Apr 22 14:47:38.943445 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.943398 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-podres\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:38.943646 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.943470 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-lib-modules\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:38.943646 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.943558 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-proc\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:38.943646 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.943618 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-sys\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:38.943816 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:38.943650 2574 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwdsm\" (UniqueName: \"kubernetes.io/projected/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-kube-api-access-hwdsm\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045036 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.044927 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-lib-modules\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045036 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045009 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-proc\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045055 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-sys\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045083 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hwdsm\" (UniqueName: \"kubernetes.io/projected/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-kube-api-access-hwdsm\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045099 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-lib-modules\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045127 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-proc\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045130 2574 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-podres\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045144 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-sys\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.045263 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.045215 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-podres\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.054500 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.054472 2574 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwdsm\" (UniqueName: \"kubernetes.io/projected/736f4b3c-ce41-44b2-8d33-6b57c7d5c697-kube-api-access-hwdsm\") pod \"perf-node-gather-daemonset-z8jhz\" (UID: \"736f4b3c-ce41-44b2-8d33-6b57c7d5c697\") " pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.064359 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.064342 2574 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.192210 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.192176 2574 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz"] Apr 22 14:47:39.195326 ip-10-0-136-18 kubenswrapper[2574]: W0422 14:47:39.195288 2574 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod736f4b3c_ce41_44b2_8d33_6b57c7d5c697.slice/crio-d48040a7be0b43e93c865b4edeba05893342483884f4285c0f6f7a93c604b6ab WatchSource:0}: Error finding container d48040a7be0b43e93c865b4edeba05893342483884f4285c0f6f7a93c604b6ab: Status 404 returned error can't find the container with id d48040a7be0b43e93c865b4edeba05893342483884f4285c0f6f7a93c604b6ab Apr 22 14:47:39.196837 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.196821 2574 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 14:47:39.578928 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.578839 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-4kpkf_0ef3c015-11a0-43bf-b3db-a4ee755ab04b/dns/0.log" Apr 22 14:47:39.598570 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.598538 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-4kpkf_0ef3c015-11a0-43bf-b3db-a4ee755ab04b/kube-rbac-proxy/0.log" Apr 22 14:47:39.648725 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.648661 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" event={"ID":"736f4b3c-ce41-44b2-8d33-6b57c7d5c697","Type":"ContainerStarted","Data":"f371c2c71f330bd6180b2cc3f746dd2ec870ad315d7f3305ca1990a6810dd44a"} Apr 22 14:47:39.648923 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.648734 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:39.648923 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.648746 2574 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" event={"ID":"736f4b3c-ce41-44b2-8d33-6b57c7d5c697","Type":"ContainerStarted","Data":"d48040a7be0b43e93c865b4edeba05893342483884f4285c0f6f7a93c604b6ab"} Apr 22 14:47:39.665725 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.665642 2574 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" podStartSLOduration=1.6656265239999999 podStartE2EDuration="1.665626524s" podCreationTimestamp="2026-04-22 14:47:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:47:39.663220514 +0000 UTC m=+1973.131098244" watchObservedRunningTime="2026-04-22 14:47:39.665626524 +0000 UTC m=+1973.133504253" Apr 22 14:47:39.708257 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:39.708223 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-d5pcf_cd5f393a-8e7d-4fec-ba9c-362d5579a518/dns-node-resolver/0.log" Apr 22 14:47:40.326707 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:40.326661 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-dx5kd_3688a996-ab5a-4d60-b8e0-19b8a75bae48/node-ca/0.log" Apr 22 14:47:41.207443 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:41.207404 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_istiod-openshift-gateway-55ff986f96-7cccv_a2ec9ada-974f-4fa8-80fe-f7a1798b0fff/discovery/0.log" Apr 22 14:47:41.675637 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:41.675608 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-b6brs_daaf870a-47fd-4fba-8f87-375c235c1a4c/serve-healthcheck-canary/0.log" Apr 22 14:47:42.133146 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:42.133032 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-zkfsf_727af150-86df-4722-a839-797b4cb1db69/insights-operator/1.log" Apr 22 14:47:42.133696 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:42.133648 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-zkfsf_727af150-86df-4722-a839-797b4cb1db69/insights-operator/0.log" Apr 22 14:47:42.156060 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:42.156029 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-2k8nd_3dffc553-a2b4-4156-af72-d9a25248a386/kube-rbac-proxy/0.log" Apr 22 14:47:42.177258 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:42.177224 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-2k8nd_3dffc553-a2b4-4156-af72-d9a25248a386/exporter/0.log" Apr 22 14:47:42.200997 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:42.200970 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-2k8nd_3dffc553-a2b4-4156-af72-d9a25248a386/extractor/0.log" Apr 22 14:47:44.837195 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:44.837167 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-lws-operator_lws-controller-manager-879f8864c-95qvc_10673011-5fdc-4c8e-aa42-ce89381bc7f6/manager/0.log" Apr 22 14:47:45.431153 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:45.431101 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve_llmisvc-controller-manager-6649659bc5-p7vvj_9fd883d2-6b84-4233-9b58-e0bda80ad8cc/manager/0.log" Apr 22 14:47:45.663932 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:45.663907 2574 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-54rh7/perf-node-gather-daemonset-z8jhz" Apr 22 14:47:50.700553 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:50.700518 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-nqqx6_c1d8e35d-172e-4b63-9962-341853586a26/migrator/0.log" Apr 22 14:47:50.726377 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:50.726347 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-nqqx6_c1d8e35d-172e-4b63-9962-341853586a26/graceful-termination/0.log" Apr 22 14:47:51.081431 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:51.081340 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-9lcw9_36bf42f9-41dd-4d29-9f4c-2472e7df5572/kube-storage-version-migrator-operator/1.log" Apr 22 14:47:51.083235 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:51.083206 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-9lcw9_36bf42f9-41dd-4d29-9f4c-2472e7df5572/kube-storage-version-migrator-operator/0.log" Apr 22 14:47:52.087998 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.087922 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-25xzb_cc6913ef-cc9f-4a41-971f-4e5dda9212fd/kube-multus-additional-cni-plugins/0.log" Apr 22 14:47:52.106963 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.106939 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-25xzb_cc6913ef-cc9f-4a41-971f-4e5dda9212fd/egress-router-binary-copy/0.log" Apr 22 14:47:52.125827 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.125801 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-25xzb_cc6913ef-cc9f-4a41-971f-4e5dda9212fd/cni-plugins/0.log" Apr 22 14:47:52.144097 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.144062 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-25xzb_cc6913ef-cc9f-4a41-971f-4e5dda9212fd/bond-cni-plugin/0.log" Apr 22 14:47:52.163480 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.163443 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-25xzb_cc6913ef-cc9f-4a41-971f-4e5dda9212fd/routeoverride-cni/0.log" Apr 22 14:47:52.183839 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.183810 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-25xzb_cc6913ef-cc9f-4a41-971f-4e5dda9212fd/whereabouts-cni-bincopy/0.log" Apr 22 14:47:52.202122 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.202092 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-25xzb_cc6913ef-cc9f-4a41-971f-4e5dda9212fd/whereabouts-cni/0.log" Apr 22 14:47:52.652012 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.651978 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hwrgf_4ec55872-0394-47c2-9108-d21ab221b5f8/kube-multus/0.log" Apr 22 14:47:52.794359 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.794327 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-p65qg_1eca98e2-015f-4ce7-bcd6-76878fa26d4b/network-metrics-daemon/0.log" Apr 22 14:47:52.815174 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:52.815128 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-p65qg_1eca98e2-015f-4ce7-bcd6-76878fa26d4b/kube-rbac-proxy/0.log" Apr 22 14:47:53.981545 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:53.981494 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-controller/0.log" Apr 22 14:47:53.998392 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:53.998367 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/0.log" Apr 22 14:47:54.016008 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:54.015984 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovn-acl-logging/1.log" Apr 22 14:47:54.042302 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:54.042279 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/kube-rbac-proxy-node/0.log" Apr 22 14:47:54.065276 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:54.065257 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/kube-rbac-proxy-ovn-metrics/0.log" Apr 22 14:47:54.083345 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:54.083298 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/northd/0.log" Apr 22 14:47:54.109918 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:54.109898 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/nbdb/0.log" Apr 22 14:47:54.197640 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:54.197614 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/sbdb/0.log" Apr 22 14:47:54.376535 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:54.376453 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gpjz8_3cff0f27-e53b-4843-a2ed-7e885a500c4a/ovnkube-controller/0.log" Apr 22 14:47:55.777502 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:55.777469 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-8894fc9bd-9m6vn_1bdda416-04a6-4a10-a11e-ead137503dd2/check-endpoints/0.log" Apr 22 14:47:55.799230 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:55.799207 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-68xhd_f67ee112-d296-44c7-9c63-b48da0c3bc44/network-check-target-container/0.log" Apr 22 14:47:56.831411 ip-10-0-136-18 kubenswrapper[2574]: I0422 14:47:56.831382 2574 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-9dhqr_6a01b9d7-2c6c-4443-a988-d06252bd1a67/iptables-alerter/0.log"