Apr 22 13:18:41.797487 ip-10-0-128-54 systemd[1]: kubelet.service: Failed to load environment files: No such file or directory Apr 22 13:18:41.797499 ip-10-0-128-54 systemd[1]: kubelet.service: Failed to run 'start-pre' task: No such file or directory Apr 22 13:18:41.797506 ip-10-0-128-54 systemd[1]: kubelet.service: Failed with result 'resources'. Apr 22 13:18:41.797740 ip-10-0-128-54 systemd[1]: Failed to start Kubernetes Kubelet. Apr 22 13:18:52.935052 ip-10-0-128-54 systemd[1]: kubelet.service: Failed to schedule restart job: Unit crio.service not found. Apr 22 13:18:52.935067 ip-10-0-128-54 systemd[1]: kubelet.service: Failed with result 'resources'. -- Boot ca21cd44bb4444148c0d3009e338dcc6 -- Apr 22 13:21:21.958747 ip-10-0-128-54 systemd[1]: Starting Kubernetes Kubelet... Apr 22 13:21:22.363134 ip-10-0-128-54 kubenswrapper[2579]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 13:21:22.363134 ip-10-0-128-54 kubenswrapper[2579]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 22 13:21:22.363134 ip-10-0-128-54 kubenswrapper[2579]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 13:21:22.363134 ip-10-0-128-54 kubenswrapper[2579]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 22 13:21:22.363134 ip-10-0-128-54 kubenswrapper[2579]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 13:21:22.366140 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.366018 2579 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 22 13:21:22.369685 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369653 2579 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 13:21:22.369685 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369682 2579 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 13:21:22.369685 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369687 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 13:21:22.369685 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369691 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369694 2579 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369698 2579 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369700 2579 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369705 2579 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369709 2579 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369713 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369716 2579 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369719 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369723 2579 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369725 2579 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369728 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369731 2579 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369733 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369736 2579 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369739 2579 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369742 2579 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369745 2579 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369747 2579 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369750 2579 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 13:21:22.369826 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369753 2579 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369755 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369759 2579 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369761 2579 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369765 2579 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369767 2579 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369770 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369772 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369775 2579 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369778 2579 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369780 2579 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369783 2579 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369785 2579 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369788 2579 feature_gate.go:328] unrecognized feature gate: Example Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369792 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369795 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369798 2579 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369802 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369806 2579 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369810 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 13:21:22.370316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369814 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369818 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369821 2579 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369826 2579 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369831 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369834 2579 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369837 2579 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369840 2579 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369843 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369846 2579 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369849 2579 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369852 2579 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369855 2579 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369859 2579 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369863 2579 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369867 2579 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369870 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369872 2579 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369875 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 13:21:22.370851 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369877 2579 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369880 2579 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369883 2579 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369885 2579 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369889 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369892 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369895 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369899 2579 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369902 2579 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369905 2579 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369908 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369910 2579 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369913 2579 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369916 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369918 2579 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369921 2579 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369923 2579 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369926 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369929 2579 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 13:21:22.371355 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369931 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369935 2579 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369938 2579 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369942 2579 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.369944 2579 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371624 2579 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371634 2579 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371637 2579 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371641 2579 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371644 2579 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371647 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371650 2579 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371653 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371656 2579 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371659 2579 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371662 2579 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371664 2579 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371668 2579 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371671 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371674 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 13:21:22.371853 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371677 2579 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371679 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371682 2579 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371684 2579 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371687 2579 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371689 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371692 2579 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371694 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371697 2579 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371699 2579 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371702 2579 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371704 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371707 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371709 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371713 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371716 2579 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371718 2579 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371721 2579 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371723 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371726 2579 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 13:21:22.372356 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371728 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371731 2579 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371735 2579 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371738 2579 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371741 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371744 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371747 2579 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371750 2579 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371753 2579 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371756 2579 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371765 2579 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371769 2579 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371771 2579 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371774 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371778 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371780 2579 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371783 2579 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371786 2579 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371788 2579 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 13:21:22.372897 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371791 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371794 2579 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371796 2579 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371799 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371802 2579 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371804 2579 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371808 2579 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371813 2579 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371816 2579 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371819 2579 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371823 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371826 2579 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371828 2579 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371831 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371834 2579 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371837 2579 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371839 2579 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371842 2579 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371844 2579 feature_gate.go:328] unrecognized feature gate: Example Apr 22 13:21:22.373389 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371847 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371849 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371853 2579 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371855 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371858 2579 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371862 2579 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371865 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371867 2579 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371870 2579 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371873 2579 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371875 2579 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371878 2579 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.371880 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.371979 2579 flags.go:64] FLAG: --address="0.0.0.0" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.371988 2579 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.371995 2579 flags.go:64] FLAG: --anonymous-auth="true" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372000 2579 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372006 2579 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372010 2579 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372015 2579 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372020 2579 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 22 13:21:22.373896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372024 2579 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372027 2579 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372031 2579 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372035 2579 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372038 2579 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372041 2579 flags.go:64] FLAG: --cgroup-root="" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372044 2579 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372047 2579 flags.go:64] FLAG: --client-ca-file="" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372050 2579 flags.go:64] FLAG: --cloud-config="" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372053 2579 flags.go:64] FLAG: --cloud-provider="external" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372056 2579 flags.go:64] FLAG: --cluster-dns="[]" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372061 2579 flags.go:64] FLAG: --cluster-domain="" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372064 2579 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372068 2579 flags.go:64] FLAG: --config-dir="" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372071 2579 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372074 2579 flags.go:64] FLAG: --container-log-max-files="5" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372079 2579 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372083 2579 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372086 2579 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372090 2579 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372093 2579 flags.go:64] FLAG: --contention-profiling="false" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372096 2579 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372099 2579 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372102 2579 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372105 2579 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 22 13:21:22.374516 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372110 2579 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372113 2579 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372116 2579 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372121 2579 flags.go:64] FLAG: --enable-load-reader="false" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372125 2579 flags.go:64] FLAG: --enable-server="true" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372128 2579 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372133 2579 flags.go:64] FLAG: --event-burst="100" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372136 2579 flags.go:64] FLAG: --event-qps="50" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372139 2579 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372143 2579 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372146 2579 flags.go:64] FLAG: --eviction-hard="" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372150 2579 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372153 2579 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372156 2579 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372159 2579 flags.go:64] FLAG: --eviction-soft="" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372162 2579 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372166 2579 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372169 2579 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372172 2579 flags.go:64] FLAG: --experimental-mounter-path="" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372175 2579 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372178 2579 flags.go:64] FLAG: --fail-swap-on="true" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372181 2579 flags.go:64] FLAG: --feature-gates="" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372185 2579 flags.go:64] FLAG: --file-check-frequency="20s" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372189 2579 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372193 2579 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 22 13:21:22.375136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372198 2579 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372201 2579 flags.go:64] FLAG: --healthz-port="10248" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372204 2579 flags.go:64] FLAG: --help="false" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372207 2579 flags.go:64] FLAG: --hostname-override="ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372211 2579 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372214 2579 flags.go:64] FLAG: --http-check-frequency="20s" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372217 2579 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372220 2579 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372224 2579 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372227 2579 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372230 2579 flags.go:64] FLAG: --image-service-endpoint="" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372233 2579 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372236 2579 flags.go:64] FLAG: --kube-api-burst="100" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372239 2579 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372243 2579 flags.go:64] FLAG: --kube-api-qps="50" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372247 2579 flags.go:64] FLAG: --kube-reserved="" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372250 2579 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372253 2579 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372256 2579 flags.go:64] FLAG: --kubelet-cgroups="" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372261 2579 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372263 2579 flags.go:64] FLAG: --lock-file="" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372266 2579 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372270 2579 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372273 2579 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 22 13:21:22.375822 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372280 2579 flags.go:64] FLAG: --log-json-split-stream="false" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372283 2579 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372287 2579 flags.go:64] FLAG: --log-text-split-stream="false" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372290 2579 flags.go:64] FLAG: --logging-format="text" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372294 2579 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372297 2579 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372301 2579 flags.go:64] FLAG: --manifest-url="" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372304 2579 flags.go:64] FLAG: --manifest-url-header="" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372309 2579 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372321 2579 flags.go:64] FLAG: --max-open-files="1000000" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372326 2579 flags.go:64] FLAG: --max-pods="110" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372330 2579 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372333 2579 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372336 2579 flags.go:64] FLAG: --memory-manager-policy="None" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372339 2579 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372343 2579 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372346 2579 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372350 2579 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372361 2579 flags.go:64] FLAG: --node-status-max-images="50" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372364 2579 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372368 2579 flags.go:64] FLAG: --oom-score-adj="-999" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372372 2579 flags.go:64] FLAG: --pod-cidr="" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372375 2579 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 22 13:21:22.376408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372383 2579 flags.go:64] FLAG: --pod-manifest-path="" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372386 2579 flags.go:64] FLAG: --pod-max-pids="-1" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372389 2579 flags.go:64] FLAG: --pods-per-core="0" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372393 2579 flags.go:64] FLAG: --port="10250" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372397 2579 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372400 2579 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-02e454076620de9ba" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372404 2579 flags.go:64] FLAG: --qos-reserved="" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372407 2579 flags.go:64] FLAG: --read-only-port="10255" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372410 2579 flags.go:64] FLAG: --register-node="true" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372420 2579 flags.go:64] FLAG: --register-schedulable="true" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372423 2579 flags.go:64] FLAG: --register-with-taints="" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372451 2579 flags.go:64] FLAG: --registry-burst="10" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372454 2579 flags.go:64] FLAG: --registry-qps="5" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372457 2579 flags.go:64] FLAG: --reserved-cpus="" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372461 2579 flags.go:64] FLAG: --reserved-memory="" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372465 2579 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372468 2579 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372472 2579 flags.go:64] FLAG: --rotate-certificates="false" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372475 2579 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372485 2579 flags.go:64] FLAG: --runonce="false" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372488 2579 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372492 2579 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372495 2579 flags.go:64] FLAG: --seccomp-default="false" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372498 2579 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372502 2579 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372505 2579 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 22 13:21:22.377086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372508 2579 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372512 2579 flags.go:64] FLAG: --storage-driver-password="root" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372515 2579 flags.go:64] FLAG: --storage-driver-secure="false" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372518 2579 flags.go:64] FLAG: --storage-driver-table="stats" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372521 2579 flags.go:64] FLAG: --storage-driver-user="root" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372525 2579 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372528 2579 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372532 2579 flags.go:64] FLAG: --system-cgroups="" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372535 2579 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372545 2579 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372548 2579 flags.go:64] FLAG: --tls-cert-file="" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372552 2579 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372561 2579 flags.go:64] FLAG: --tls-min-version="" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372564 2579 flags.go:64] FLAG: --tls-private-key-file="" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372567 2579 flags.go:64] FLAG: --topology-manager-policy="none" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372572 2579 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372575 2579 flags.go:64] FLAG: --topology-manager-scope="container" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372578 2579 flags.go:64] FLAG: --v="2" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372584 2579 flags.go:64] FLAG: --version="false" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372589 2579 flags.go:64] FLAG: --vmodule="" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372594 2579 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.372597 2579 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372733 2579 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372738 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 13:21:22.377741 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372742 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372745 2579 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372751 2579 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372754 2579 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372758 2579 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372761 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372764 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372766 2579 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372769 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372772 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372775 2579 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372778 2579 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372781 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372784 2579 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372786 2579 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372789 2579 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372792 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372796 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372799 2579 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372801 2579 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 13:21:22.378316 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372804 2579 feature_gate.go:328] unrecognized feature gate: Example Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372807 2579 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372809 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372815 2579 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372818 2579 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372820 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372823 2579 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372826 2579 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372828 2579 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372831 2579 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372834 2579 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372836 2579 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372839 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372842 2579 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372844 2579 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372851 2579 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372855 2579 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372858 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372861 2579 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 13:21:22.378842 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372864 2579 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372867 2579 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372870 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372873 2579 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372875 2579 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372878 2579 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372880 2579 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372883 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372886 2579 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372889 2579 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372894 2579 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372896 2579 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372899 2579 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372902 2579 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372905 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372908 2579 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372912 2579 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372915 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372918 2579 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372922 2579 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372924 2579 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 13:21:22.379312 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372927 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372931 2579 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372935 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372940 2579 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372942 2579 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372946 2579 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372949 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372951 2579 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372955 2579 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372958 2579 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372960 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372963 2579 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372965 2579 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372968 2579 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372970 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372973 2579 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372975 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372978 2579 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372980 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 13:21:22.379840 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372984 2579 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372986 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372991 2579 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372994 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.372997 2579 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.373010 2579 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.379909 2579 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.379927 2579 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.379980 2579 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.379985 2579 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.379989 2579 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.379992 2579 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.379996 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.379998 2579 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380001 2579 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380003 2579 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 13:21:22.380313 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380006 2579 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380009 2579 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380012 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380014 2579 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380017 2579 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380021 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380023 2579 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380026 2579 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380029 2579 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380031 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380034 2579 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380036 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380039 2579 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380042 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380044 2579 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380047 2579 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380049 2579 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380052 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380055 2579 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380058 2579 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 13:21:22.380944 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380061 2579 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380064 2579 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380067 2579 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380070 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380073 2579 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380076 2579 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380078 2579 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380081 2579 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380084 2579 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380086 2579 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380091 2579 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380096 2579 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380100 2579 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380103 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380106 2579 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380109 2579 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380112 2579 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380114 2579 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380118 2579 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 13:21:22.381630 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380121 2579 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380124 2579 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380127 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380130 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380132 2579 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380135 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380138 2579 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380141 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380144 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380147 2579 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380150 2579 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380153 2579 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380156 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380158 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380161 2579 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380163 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380166 2579 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380169 2579 feature_gate.go:328] unrecognized feature gate: Example Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380171 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380174 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 13:21:22.382103 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380176 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380179 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380182 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380184 2579 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380187 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380189 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380192 2579 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380194 2579 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380197 2579 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380199 2579 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380202 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380204 2579 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380208 2579 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380210 2579 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380213 2579 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380215 2579 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380218 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380220 2579 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 13:21:22.382679 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380222 2579 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.380228 2579 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380844 2579 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380867 2579 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380877 2579 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380887 2579 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380895 2579 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380903 2579 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380911 2579 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380917 2579 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380925 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380932 2579 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380939 2579 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380946 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380962 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 13:21:22.383126 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380969 2579 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380977 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380984 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380990 2579 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.380997 2579 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381004 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381011 2579 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381018 2579 feature_gate.go:328] unrecognized feature gate: Example Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381025 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381033 2579 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381037 2579 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381046 2579 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381051 2579 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381056 2579 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381062 2579 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381066 2579 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381070 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381088 2579 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381092 2579 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381097 2579 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 13:21:22.383529 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381101 2579 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381105 2579 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381109 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381114 2579 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381124 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381128 2579 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381132 2579 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381136 2579 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381141 2579 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381145 2579 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381149 2579 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381153 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381158 2579 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381162 2579 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381166 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381171 2579 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381175 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381184 2579 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381189 2579 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381193 2579 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 13:21:22.384011 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381197 2579 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381202 2579 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381206 2579 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381210 2579 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381215 2579 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381220 2579 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381225 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381229 2579 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381233 2579 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381242 2579 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381246 2579 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381250 2579 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381254 2579 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381259 2579 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381264 2579 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381269 2579 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381274 2579 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381279 2579 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381284 2579 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 13:21:22.384511 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381288 2579 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381292 2579 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381297 2579 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381306 2579 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381311 2579 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381317 2579 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381323 2579 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381327 2579 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381332 2579 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381336 2579 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381340 2579 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381344 2579 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381349 2579 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:22.381353 2579 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.381362 2579 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 13:21:22.384971 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.382227 2579 server.go:962] "Client rotation is on, will bootstrap in background" Apr 22 13:21:22.385341 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.384356 2579 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 22 13:21:22.385341 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.385299 2579 server.go:1019] "Starting client certificate rotation" Apr 22 13:21:22.385412 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.385394 2579 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 13:21:22.385458 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.385454 2579 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 13:21:22.409404 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.409384 2579 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 13:21:22.413800 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.413781 2579 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 13:21:22.428632 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.428609 2579 log.go:25] "Validated CRI v1 runtime API" Apr 22 13:21:22.434057 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.434037 2579 log.go:25] "Validated CRI v1 image API" Apr 22 13:21:22.435792 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.435767 2579 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 22 13:21:22.438579 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.438558 2579 fs.go:135] Filesystem UUIDs: map[7B77-95E7:/dev/nvme0n1p2 7f9877fc-3379-48b9-937e-35653436311a:/dev/nvme0n1p4 d81d3ab6-8087-4856-aec6-b7f92dd41a35:/dev/nvme0n1p3] Apr 22 13:21:22.438670 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.438579 2579 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 22 13:21:22.441218 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.441199 2579 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 13:21:22.445264 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.445141 2579 manager.go:217] Machine: {Timestamp:2026-04-22 13:21:22.4430238 +0000 UTC m=+0.379410656 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3100757 MemoryCapacity:33164492800 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2efe5611b27860248864e04d88ea7b SystemUUID:ec2efe56-11b2-7860-2488-64e04d88ea7b BootID:ca21cd44-bb44-4414-8c0d-3009e338dcc6 Filesystems:[{Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16582246400 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16582246400 Type:vfs Inodes:4048400 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6632898560 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:4e:71:bd:86:f9 Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:4e:71:bd:86:f9 Speed:0 Mtu:9001} {Name:ovs-system MacAddress:7a:2f:e3:0f:c0:ef Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33164492800 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 22 13:21:22.445264 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.445255 2579 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 22 13:21:22.445404 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.445345 2579 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 22 13:21:22.447806 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.447783 2579 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 22 13:21:22.447937 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.447808 2579 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-128-54.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 22 13:21:22.447988 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.447946 2579 topology_manager.go:138] "Creating topology manager with none policy" Apr 22 13:21:22.447988 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.447955 2579 container_manager_linux.go:306] "Creating device plugin manager" Apr 22 13:21:22.447988 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.447973 2579 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 13:21:22.448635 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.448625 2579 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 13:21:22.449454 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.449444 2579 state_mem.go:36] "Initialized new in-memory state store" Apr 22 13:21:22.449732 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.449723 2579 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 22 13:21:22.452207 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.452197 2579 kubelet.go:491] "Attempting to sync node with API server" Apr 22 13:21:22.452247 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.452211 2579 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 22 13:21:22.452247 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.452223 2579 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 22 13:21:22.452247 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.452232 2579 kubelet.go:397] "Adding apiserver pod source" Apr 22 13:21:22.452247 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.452241 2579 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 22 13:21:22.453279 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.453262 2579 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 13:21:22.453279 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.453280 2579 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 13:21:22.455945 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.455929 2579 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 22 13:21:22.457277 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.457265 2579 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 22 13:21:22.459019 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.458998 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 22 13:21:22.459019 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459018 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459025 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459031 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459037 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459043 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459049 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459054 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459061 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459067 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459075 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 22 13:21:22.459122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459085 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 22 13:21:22.459825 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459814 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 22 13:21:22.459866 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459828 2579 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 22 13:21:22.459913 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.459898 2579 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-vv52j" Apr 22 13:21:22.463492 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.463477 2579 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 22 13:21:22.463595 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.463510 2579 server.go:1295] "Started kubelet" Apr 22 13:21:22.463692 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.463676 2579 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-128-54.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 22 13:21:22.463692 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.463656 2579 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-128-54.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 22 13:21:22.463787 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.463661 2579 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 22 13:21:22.463787 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.463723 2579 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 22 13:21:22.463851 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.463786 2579 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 22 13:21:22.463983 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.463656 2579 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 22 13:21:22.464295 ip-10-0-128-54 systemd[1]: Started Kubernetes Kubelet. Apr 22 13:21:22.464797 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.464779 2579 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 22 13:21:22.466076 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.466061 2579 server.go:317] "Adding debug handlers to kubelet server" Apr 22 13:21:22.467241 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.467223 2579 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-vv52j" Apr 22 13:21:22.468910 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.467953 2579 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-128-54.ec2.internal.18a8b0780162fcfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-128-54.ec2.internal,UID:ip-10-0-128-54.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-10-0-128-54.ec2.internal,},FirstTimestamp:2026-04-22 13:21:22.463489278 +0000 UTC m=+0.399876131,LastTimestamp:2026-04-22 13:21:22.463489278 +0000 UTC m=+0.399876131,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-128-54.ec2.internal,}" Apr 22 13:21:22.471404 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.471384 2579 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 22 13:21:22.471837 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.471808 2579 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 22 13:21:22.471931 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.471851 2579 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 22 13:21:22.472441 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472413 2579 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 22 13:21:22.472515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472445 2579 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 22 13:21:22.472515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472414 2579 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 22 13:21:22.472599 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472544 2579 factory.go:55] Registering systemd factory Apr 22 13:21:22.472599 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472561 2579 factory.go:223] Registration of the systemd container factory successfully Apr 22 13:21:22.472703 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472549 2579 reconstruct.go:97] "Volume reconstruction finished" Apr 22 13:21:22.472703 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472614 2579 reconciler.go:26] "Reconciler: start to sync state" Apr 22 13:21:22.472703 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.472633 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:22.472837 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472759 2579 factory.go:153] Registering CRI-O factory Apr 22 13:21:22.472837 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472769 2579 factory.go:223] Registration of the crio container factory successfully Apr 22 13:21:22.472837 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472819 2579 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 22 13:21:22.472979 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472842 2579 factory.go:103] Registering Raw factory Apr 22 13:21:22.472979 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.472856 2579 manager.go:1196] Started watching for new ooms in manager Apr 22 13:21:22.473357 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.473345 2579 manager.go:319] Starting recovery of all containers Apr 22 13:21:22.479145 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.478991 2579 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 13:21:22.482993 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.482966 2579 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-10-0-128-54.ec2.internal\" not found" node="ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.483920 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.483903 2579 manager.go:324] Recovery completed Apr 22 13:21:22.488218 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.488205 2579 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 13:21:22.490365 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.490350 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientMemory" Apr 22 13:21:22.490458 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.490376 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 13:21:22.490458 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.490386 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientPID" Apr 22 13:21:22.490868 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.490855 2579 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 22 13:21:22.490868 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.490866 2579 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 22 13:21:22.490939 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.490885 2579 state_mem.go:36] "Initialized new in-memory state store" Apr 22 13:21:22.493095 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.493083 2579 policy_none.go:49] "None policy: Start" Apr 22 13:21:22.493145 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.493099 2579 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 22 13:21:22.493145 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.493108 2579 state_mem.go:35] "Initializing new in-memory state store" Apr 22 13:21:22.522148 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.522127 2579 manager.go:341] "Starting Device Plugin manager" Apr 22 13:21:22.522272 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.522164 2579 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 22 13:21:22.522272 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.522175 2579 server.go:85] "Starting device plugin registration server" Apr 22 13:21:22.522484 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.522471 2579 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 22 13:21:22.522532 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.522488 2579 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 22 13:21:22.522600 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.522559 2579 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 22 13:21:22.522651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.522639 2579 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 22 13:21:22.522651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.522648 2579 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 22 13:21:22.523228 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.523206 2579 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 22 13:21:22.523353 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.523247 2579 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:22.598567 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.598521 2579 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 22 13:21:22.599792 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.599774 2579 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 22 13:21:22.599858 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.599802 2579 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 22 13:21:22.599858 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.599823 2579 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 22 13:21:22.599858 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.599829 2579 kubelet.go:2451] "Starting kubelet main sync loop" Apr 22 13:21:22.599969 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.599864 2579 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 22 13:21:22.602422 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.602402 2579 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 13:21:22.623623 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.623542 2579 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 13:21:22.624697 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.624678 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientMemory" Apr 22 13:21:22.624819 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.624709 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 13:21:22.624819 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.624720 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientPID" Apr 22 13:21:22.624819 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.624744 2579 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.631376 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.631357 2579 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.631496 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.631384 2579 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-128-54.ec2.internal\": node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:22.643294 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.643271 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:22.700969 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.700921 2579 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal"] Apr 22 13:21:22.701080 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.701024 2579 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 13:21:22.701994 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.701977 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientMemory" Apr 22 13:21:22.702111 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.702013 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 13:21:22.702111 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.702027 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientPID" Apr 22 13:21:22.703286 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.703269 2579 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 13:21:22.703419 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.703405 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.703485 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.703445 2579 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 13:21:22.704634 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.704620 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientMemory" Apr 22 13:21:22.704728 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.704647 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 13:21:22.704728 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.704661 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientPID" Apr 22 13:21:22.704728 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.704703 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientMemory" Apr 22 13:21:22.704728 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.704723 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 13:21:22.704897 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.704734 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientPID" Apr 22 13:21:22.705764 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.705749 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.705842 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.705778 2579 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 13:21:22.706525 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.706501 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientMemory" Apr 22 13:21:22.706612 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.706530 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 13:21:22.706612 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.706543 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeHasSufficientPID" Apr 22 13:21:22.739494 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.739464 2579 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-128-54.ec2.internal\" not found" node="ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.743915 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.743897 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:22.744015 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.743988 2579 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-128-54.ec2.internal\" not found" node="ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.774351 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.774325 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/567021ad3684c4b7c9c2e8061f7ecdc6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal\" (UID: \"567021ad3684c4b7c9c2e8061f7ecdc6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.774351 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.774354 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/16c68cd8f43f9f069938f5a899c05cc7-config\") pod \"kube-apiserver-proxy-ip-10-0-128-54.ec2.internal\" (UID: \"16c68cd8f43f9f069938f5a899c05cc7\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.774538 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.774372 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/567021ad3684c4b7c9c2e8061f7ecdc6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal\" (UID: \"567021ad3684c4b7c9c2e8061f7ecdc6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.844584 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.844556 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:22.874963 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.874902 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/567021ad3684c4b7c9c2e8061f7ecdc6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal\" (UID: \"567021ad3684c4b7c9c2e8061f7ecdc6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.874963 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.874933 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/567021ad3684c4b7c9c2e8061f7ecdc6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal\" (UID: \"567021ad3684c4b7c9c2e8061f7ecdc6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.874963 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.874949 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/16c68cd8f43f9f069938f5a899c05cc7-config\") pod \"kube-apiserver-proxy-ip-10-0-128-54.ec2.internal\" (UID: \"16c68cd8f43f9f069938f5a899c05cc7\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.875091 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.874991 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/567021ad3684c4b7c9c2e8061f7ecdc6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal\" (UID: \"567021ad3684c4b7c9c2e8061f7ecdc6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.875091 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.874994 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/567021ad3684c4b7c9c2e8061f7ecdc6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal\" (UID: \"567021ad3684c4b7c9c2e8061f7ecdc6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.875091 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:22.875029 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/16c68cd8f43f9f069938f5a899c05cc7-config\") pod \"kube-apiserver-proxy-ip-10-0-128-54.ec2.internal\" (UID: \"16c68cd8f43f9f069938f5a899c05cc7\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" Apr 22 13:21:22.945237 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:22.945200 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.041932 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.041906 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:23.045459 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.045425 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.046550 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.046534 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" Apr 22 13:21:23.145654 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.145567 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.246089 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.246058 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.346591 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.346558 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.385059 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.385025 2579 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 22 13:21:23.385672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.385174 2579 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 22 13:21:23.385672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.385205 2579 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 22 13:21:23.447400 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.447329 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.469652 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.469611 2579 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-21 13:16:22 +0000 UTC" deadline="2027-11-20 03:14:34.90865815 +0000 UTC" Apr 22 13:21:23.469652 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.469646 2579 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="13837h53m11.439015729s" Apr 22 13:21:23.472067 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.472033 2579 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 22 13:21:23.486833 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.486798 2579 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 13:21:23.488224 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.488198 2579 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 13:21:23.505417 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.505379 2579 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-zm75g" Apr 22 13:21:23.512742 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.512714 2579 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-zm75g" Apr 22 13:21:23.548176 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.548142 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.570495 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:23.570458 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod567021ad3684c4b7c9c2e8061f7ecdc6.slice/crio-cf4f5c83cd4c5a3dda365d8dfaea2c77e97042e3c49c2b115e2a8111497e8cd0 WatchSource:0}: Error finding container cf4f5c83cd4c5a3dda365d8dfaea2c77e97042e3c49c2b115e2a8111497e8cd0: Status 404 returned error can't find the container with id cf4f5c83cd4c5a3dda365d8dfaea2c77e97042e3c49c2b115e2a8111497e8cd0 Apr 22 13:21:23.570711 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:23.570690 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16c68cd8f43f9f069938f5a899c05cc7.slice/crio-3f0f29b9668024e6c1e8dc64e4b1e52cd5b4290b309c7612d6623149fdcc2fc7 WatchSource:0}: Error finding container 3f0f29b9668024e6c1e8dc64e4b1e52cd5b4290b309c7612d6623149fdcc2fc7: Status 404 returned error can't find the container with id 3f0f29b9668024e6c1e8dc64e4b1e52cd5b4290b309c7612d6623149fdcc2fc7 Apr 22 13:21:23.574162 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.574142 2579 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 13:21:23.602963 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.602911 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" event={"ID":"16c68cd8f43f9f069938f5a899c05cc7","Type":"ContainerStarted","Data":"3f0f29b9668024e6c1e8dc64e4b1e52cd5b4290b309c7612d6623149fdcc2fc7"} Apr 22 13:21:23.603797 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.603774 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" event={"ID":"567021ad3684c4b7c9c2e8061f7ecdc6","Type":"ContainerStarted","Data":"cf4f5c83cd4c5a3dda365d8dfaea2c77e97042e3c49c2b115e2a8111497e8cd0"} Apr 22 13:21:23.648989 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.648959 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.749483 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.749415 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.849917 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:23.849896 2579 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-128-54.ec2.internal\" not found" Apr 22 13:21:23.945843 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.945812 2579 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 13:21:23.972941 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.972910 2579 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" Apr 22 13:21:23.984295 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.984263 2579 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 13:21:23.985168 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.985154 2579 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" Apr 22 13:21:23.997254 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:23.997226 2579 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 13:21:24.215012 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.214931 2579 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 13:21:24.266704 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.266622 2579 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 13:21:24.453135 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.453098 2579 apiserver.go:52] "Watching apiserver" Apr 22 13:21:24.462070 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.462034 2579 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 22 13:21:24.463873 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.463682 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vzwxc","openshift-network-diagnostics/network-check-target-tmp8r","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss","openshift-multus/multus-grkk2","openshift-multus/network-metrics-daemon-8vcg7","openshift-network-operator/iptables-alerter-hv7gn","openshift-ovn-kubernetes/ovnkube-node-lqksf","kube-system/global-pull-secret-syncer-c7j5x","kube-system/konnectivity-agent-hrtnk","kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal","openshift-cluster-node-tuning-operator/tuned-wq59k","openshift-image-registry/node-ca-r48qq","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal"] Apr 22 13:21:24.466550 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.466455 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.467885 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.467818 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:24.468047 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.467923 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:24.469306 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.469132 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 22 13:21:24.469306 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.469161 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 22 13:21:24.469306 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.469207 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 22 13:21:24.469306 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.469265 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 22 13:21:24.470247 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.470226 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 22 13:21:24.470449 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.470417 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-m7dq6\"" Apr 22 13:21:24.470863 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.470627 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 22 13:21:24.470863 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.470744 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.474510 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.473188 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-jgsg7\"" Apr 22 13:21:24.474510 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.473831 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 22 13:21:24.474698 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.474415 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 22 13:21:24.474698 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.474530 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 22 13:21:24.474801 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.474671 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.474969 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.474952 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:24.475173 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.475146 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:24.476668 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.476646 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.477423 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.477401 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 22 13:21:24.477579 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.477474 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 22 13:21:24.477579 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.477529 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-ks7c6\"" Apr 22 13:21:24.477685 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.477662 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 22 13:21:24.477942 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.477921 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 22 13:21:24.478470 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.478449 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.479015 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.478996 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.480724 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.480758 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-xpx5q\"" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.480846 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-dqq7h\"" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.480954 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.480987 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.481127 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.481248 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.481642 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.481361 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.483465 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483446 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 22 13:21:24.483544 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483449 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-r7jpz\"" Apr 22 13:21:24.483544 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483520 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-tk9vn\"" Apr 22 13:21:24.483766 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483752 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 22 13:21:24.483878 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483858 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-systemd-units\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.483941 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483892 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-etc-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.483941 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483919 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-device-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.484040 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483943 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-etc-selinux\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.484040 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.483994 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-sys-fs\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.484140 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484048 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-cni-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.484140 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484078 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-os-release\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.484140 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484116 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:24.484350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484153 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-run-netns\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484186 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-var-lib-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484245 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-node-log\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484276 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-cni-bin\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484305 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/782146a0-22aa-427e-aeb7-43bb6980c291-ovn-node-metrics-cert\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484331 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-cnibin\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484354 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-netns\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484378 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-cni-multus\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484404 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484413 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484444 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484471 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p96m\" (UniqueName: \"kubernetes.io/projected/782146a0-22aa-427e-aeb7-43bb6980c291-kube-api-access-2p96m\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484496 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-daemon-config\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484533 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-kubelet\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484599 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-ovn\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484627 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 22 13:21:24.484651 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484644 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-cni-netd\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484672 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-env-overrides\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484697 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-ovnkube-script-lib\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484722 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-registration-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484747 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bps6g\" (UniqueName: \"kubernetes.io/projected/5dec8b65-6274-44b1-a169-ad20045acd1a-kube-api-access-bps6g\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484803 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484838 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-kubelet-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484863 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-conf-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484886 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-multus-certs\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484910 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh84l\" (UniqueName: \"kubernetes.io/projected/a4457d5d-3460-41bb-a335-8f9d5e2a4197-kube-api-access-fh84l\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484942 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-log-socket\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484965 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-socket-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.484989 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm4t7\" (UniqueName: \"kubernetes.io/projected/ea7df7e2-bf24-4c31-92ce-352960562242-kube-api-access-mm4t7\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485039 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-k8s-cni-cncf-io\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485072 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-etc-kubernetes\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485098 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-slash\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485121 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a4457d5d-3460-41bb-a335-8f9d5e2a4197-cni-binary-copy\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485145 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-socket-dir-parent\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485169 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-kubelet\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485199 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-hostroot\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485224 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-systemd\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485249 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-ovnkube-config\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485277 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485300 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-system-cni-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485358 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-cni-bin\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485630 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.485663 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.485862 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.485705 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:24.487992 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.487829 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 22 13:21:24.487992 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.487834 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 22 13:21:24.488136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.488087 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-64hkt\"" Apr 22 13:21:24.488188 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.488132 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 22 13:21:24.513446 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.513381 2579 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 13:16:23 +0000 UTC" deadline="2028-02-06 17:48:11.746728846 +0000 UTC" Apr 22 13:21:24.513446 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.513413 2579 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="15724h26m47.233320114s" Apr 22 13:21:24.573603 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.573568 2579 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 22 13:21:24.585963 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.585927 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2p96m\" (UniqueName: \"kubernetes.io/projected/782146a0-22aa-427e-aeb7-43bb6980c291-kube-api-access-2p96m\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.586134 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586056 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-sys\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.586134 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586090 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-system-cni-dir\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.586134 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586114 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.586312 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586139 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-env-overrides\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.586312 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586196 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-ovnkube-script-lib\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.586312 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586239 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-registration-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586318 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-registration-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586324 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bps6g\" (UniqueName: \"kubernetes.io/projected/5dec8b65-6274-44b1-a169-ad20045acd1a-kube-api-access-bps6g\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586368 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586396 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-kubelet-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586418 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586422 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-multus-certs\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586485 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-kubelet-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586503 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-lib-modules\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586536 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-multus-certs\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.586543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586547 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-var-lib-kubelet\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586569 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b26247c-8561-4d36-a271-45bd1c30c698-host\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586608 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-k8s-cni-cncf-io\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586642 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/6560c958-0928-4fa1-8691-8fe812c8d951-kubelet-config\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586650 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-k8s-cni-cncf-io\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586669 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-slash\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586706 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-kubelet\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586732 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-slash\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586775 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-hostroot\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586733 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-hostroot\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586767 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-kubelet\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586817 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cni-binary-copy\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586846 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-ovnkube-config\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586872 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586891 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-env-overrides\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586951 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-system-cni-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586898 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-system-cni-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.586987 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-cni-bin\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587058 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587012 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-etc-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587061 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-etc-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587065 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-cni-bin\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587088 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587116 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587141 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-node-log\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587167 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-cnibin\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.587179 2579 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587208 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-node-log\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587222 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-cnibin\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587255 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-ovnkube-script-lib\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.587267 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs podName:5dec8b65-6274-44b1-a169-ad20045acd1a nodeName:}" failed. No retries permitted until 2026-04-22 13:21:25.087226669 +0000 UTC m=+3.023613514 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs") pod "network-metrics-daemon-8vcg7" (UID: "5dec8b65-6274-44b1-a169-ad20045acd1a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587291 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-tuned\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587326 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-os-release\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587356 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587385 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/782146a0-22aa-427e-aeb7-43bb6980c291-ovnkube-config\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587389 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.587919 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587449 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587478 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-daemon-config\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587505 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-run\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587528 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-kubelet\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587552 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-ovn\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587602 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-ovn\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587602 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-kubelet\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587645 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-cni-netd\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587681 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysctl-conf\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587728 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-systemd\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587773 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-cni-netd\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587810 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2xf4\" (UniqueName: \"kubernetes.io/projected/ee95ba9f-567a-461d-88c1-d3a904c366ad-kube-api-access-c2xf4\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587842 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/6560c958-0928-4fa1-8691-8fe812c8d951-dbus\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587875 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-conf-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587910 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fh84l\" (UniqueName: \"kubernetes.io/projected/a4457d5d-3460-41bb-a335-8f9d5e2a4197-kube-api-access-fh84l\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587938 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-conf-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.587979 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f757ef9d-b853-489b-9d90-79295660b525-host-slash\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.588727 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588003 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-daemon-config\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588009 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588039 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hfgz\" (UniqueName: \"kubernetes.io/projected/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-kube-api-access-8hfgz\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588060 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-log-socket\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588084 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-socket-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588092 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-log-socket\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588113 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mm4t7\" (UniqueName: \"kubernetes.io/projected/ea7df7e2-bf24-4c31-92ce-352960562242-kube-api-access-mm4t7\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588137 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-etc-kubernetes\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588174 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-etc-kubernetes\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588188 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-socket-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588270 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpqg7\" (UniqueName: \"kubernetes.io/projected/f757ef9d-b853-489b-9d90-79295660b525-kube-api-access-xpqg7\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588292 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysconfig\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588316 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2b26247c-8561-4d36-a271-45bd1c30c698-serviceca\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588338 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a4457d5d-3460-41bb-a335-8f9d5e2a4197-cni-binary-copy\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588363 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-socket-dir-parent\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588459 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-kubernetes\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588479 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-socket-dir-parent\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.589468 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588488 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/87ad5783-96fb-40e6-abe4-0fda9e9532b6-agent-certs\") pod \"konnectivity-agent-hrtnk\" (UID: \"87ad5783-96fb-40e6-abe4-0fda9e9532b6\") " pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588515 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-systemd\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588538 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysctl-d\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588562 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-host\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588586 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cnibin\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588605 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-run-systemd\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588637 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkrn6\" (UniqueName: \"kubernetes.io/projected/2b26247c-8561-4d36-a271-45bd1c30c698-kube-api-access-hkrn6\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588667 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-systemd-units\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588711 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-device-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588738 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-etc-selinux\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588764 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-sys-fs\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588712 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-systemd-units\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588790 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-cni-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588815 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-os-release\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588823 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-sys-fs\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588839 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ee95ba9f-567a-461d-88c1-d3a904c366ad-tmp\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588794 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a4457d5d-3460-41bb-a335-8f9d5e2a4197-cni-binary-copy\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.590073 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588845 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-multus-cni-dir\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588770 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-device-dir\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588870 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/87ad5783-96fb-40e6-abe4-0fda9e9532b6-konnectivity-ca\") pod \"konnectivity-agent-hrtnk\" (UID: \"87ad5783-96fb-40e6-abe4-0fda9e9532b6\") " pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588895 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/ea7df7e2-bf24-4c31-92ce-352960562242-etc-selinux\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588903 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/f757ef9d-b853-489b-9d90-79295660b525-iptables-alerter-script\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588923 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-os-release\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588937 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-modprobe-d\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588968 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-run-netns\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.588994 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-var-lib-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589016 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-run-netns\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589021 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-cni-bin\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589042 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-var-lib-openvswitch\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589047 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/782146a0-22aa-427e-aeb7-43bb6980c291-ovn-node-metrics-cert\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589066 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-cni-bin\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589072 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-netns\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589097 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-cni-multus\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589121 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.590672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589141 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-run-netns\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.591352 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589157 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a4457d5d-3460-41bb-a335-8f9d5e2a4197-host-var-lib-cni-multus\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.591352 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589209 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/782146a0-22aa-427e-aeb7-43bb6980c291-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.591352 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.589322 2579 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 22 13:21:24.592755 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.592696 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/782146a0-22aa-427e-aeb7-43bb6980c291-ovn-node-metrics-cert\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.593529 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.593190 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 13:21:24.593529 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.593218 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 13:21:24.593529 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.593234 2579 projected.go:194] Error preparing data for projected volume kube-api-access-zbhdt for pod openshift-network-diagnostics/network-check-target-tmp8r: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:24.593529 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.593307 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt podName:4829e038-c6a2-4934-8182-fae81fb059a8 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:25.093286729 +0000 UTC m=+3.029673591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-zbhdt" (UniqueName: "kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt") pod "network-check-target-tmp8r" (UID: "4829e038-c6a2-4934-8182-fae81fb059a8") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:24.594559 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.594536 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bps6g\" (UniqueName: \"kubernetes.io/projected/5dec8b65-6274-44b1-a169-ad20045acd1a-kube-api-access-bps6g\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:24.594913 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.594896 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p96m\" (UniqueName: \"kubernetes.io/projected/782146a0-22aa-427e-aeb7-43bb6980c291-kube-api-access-2p96m\") pod \"ovnkube-node-lqksf\" (UID: \"782146a0-22aa-427e-aeb7-43bb6980c291\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.599739 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.599718 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh84l\" (UniqueName: \"kubernetes.io/projected/a4457d5d-3460-41bb-a335-8f9d5e2a4197-kube-api-access-fh84l\") pod \"multus-grkk2\" (UID: \"a4457d5d-3460-41bb-a335-8f9d5e2a4197\") " pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.600129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.600105 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm4t7\" (UniqueName: \"kubernetes.io/projected/ea7df7e2-bf24-4c31-92ce-352960562242-kube-api-access-mm4t7\") pod \"aws-ebs-csi-driver-node-qz6ss\" (UID: \"ea7df7e2-bf24-4c31-92ce-352960562242\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.690045 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690005 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cni-binary-copy\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.690209 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690065 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.690209 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.690164 2579 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:24.690316 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690207 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-tuned\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.690316 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:24.690239 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret podName:6560c958-0928-4fa1-8691-8fe812c8d951 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:25.190219591 +0000 UTC m=+3.126606434 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret") pod "global-pull-secret-syncer-c7j5x" (UID: "6560c958-0928-4fa1-8691-8fe812c8d951") : object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:24.690316 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690262 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-os-release\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.690316 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690279 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.690316 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690304 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-run\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690323 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysctl-conf\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690338 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-systemd\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690344 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-os-release\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690359 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-c2xf4\" (UniqueName: \"kubernetes.io/projected/ee95ba9f-567a-461d-88c1-d3a904c366ad-kube-api-access-c2xf4\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690380 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/6560c958-0928-4fa1-8691-8fe812c8d951-dbus\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690406 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f757ef9d-b853-489b-9d90-79295660b525-host-slash\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690412 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-run\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690456 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690483 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8hfgz\" (UniqueName: \"kubernetes.io/projected/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-kube-api-access-8hfgz\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690511 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xpqg7\" (UniqueName: \"kubernetes.io/projected/f757ef9d-b853-489b-9d90-79295660b525-kube-api-access-xpqg7\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690533 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysconfig\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.690572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690555 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2b26247c-8561-4d36-a271-45bd1c30c698-serviceca\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690588 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-kubernetes\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690612 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/87ad5783-96fb-40e6-abe4-0fda9e9532b6-agent-certs\") pod \"konnectivity-agent-hrtnk\" (UID: \"87ad5783-96fb-40e6-abe4-0fda9e9532b6\") " pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690636 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysctl-d\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690657 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-host\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690679 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cnibin\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690688 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cni-binary-copy\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690704 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hkrn6\" (UniqueName: \"kubernetes.io/projected/2b26247c-8561-4d36-a271-45bd1c30c698-kube-api-access-hkrn6\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690715 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysctl-conf\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690737 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ee95ba9f-567a-461d-88c1-d3a904c366ad-tmp\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690721 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-systemd\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690746 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-kubernetes\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690775 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/87ad5783-96fb-40e6-abe4-0fda9e9532b6-konnectivity-ca\") pod \"konnectivity-agent-hrtnk\" (UID: \"87ad5783-96fb-40e6-abe4-0fda9e9532b6\") " pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690810 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/f757ef9d-b853-489b-9d90-79295660b525-iptables-alerter-script\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690840 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-modprobe-d\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690842 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690879 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-sys\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691157 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690905 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-system-cni-dir\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690932 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690967 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-lib-modules\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690992 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-var-lib-kubelet\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691018 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b26247c-8561-4d36-a271-45bd1c30c698-host\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691043 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/6560c958-0928-4fa1-8691-8fe812c8d951-kubelet-config\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691054 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691114 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2b26247c-8561-4d36-a271-45bd1c30c698-serviceca\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691129 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/6560c958-0928-4fa1-8691-8fe812c8d951-kubelet-config\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691118 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f757ef9d-b853-489b-9d90-79295660b525-host-slash\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691179 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-system-cni-dir\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691184 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-var-lib-kubelet\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691180 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/6560c958-0928-4fa1-8691-8fe812c8d951-dbus\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.690781 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysconfig\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691226 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b26247c-8561-4d36-a271-45bd1c30c698-host\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691262 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-lib-modules\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691283 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-sys\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.691907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691284 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.692520 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691319 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-cnibin\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.692520 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691343 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-sysctl-d\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.692520 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691343 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/87ad5783-96fb-40e6-abe4-0fda9e9532b6-konnectivity-ca\") pod \"konnectivity-agent-hrtnk\" (UID: \"87ad5783-96fb-40e6-abe4-0fda9e9532b6\") " pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.692520 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691383 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-host\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.692520 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691398 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-modprobe-d\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.692520 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.691777 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/f757ef9d-b853-489b-9d90-79295660b525-iptables-alerter-script\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.693649 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.693626 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ee95ba9f-567a-461d-88c1-d3a904c366ad-tmp\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.693760 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.693649 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/ee95ba9f-567a-461d-88c1-d3a904c366ad-etc-tuned\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.693948 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.693931 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/87ad5783-96fb-40e6-abe4-0fda9e9532b6-agent-certs\") pod \"konnectivity-agent-hrtnk\" (UID: \"87ad5783-96fb-40e6-abe4-0fda9e9532b6\") " pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.699002 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.698976 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hfgz\" (UniqueName: \"kubernetes.io/projected/1d77049f-f54b-436c-8e41-5e0b8d54e8bf-kube-api-access-8hfgz\") pod \"multus-additional-cni-plugins-vzwxc\" (UID: \"1d77049f-f54b-436c-8e41-5e0b8d54e8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.699100 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.699055 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpqg7\" (UniqueName: \"kubernetes.io/projected/f757ef9d-b853-489b-9d90-79295660b525-kube-api-access-xpqg7\") pod \"iptables-alerter-hv7gn\" (UID: \"f757ef9d-b853-489b-9d90-79295660b525\") " pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.699779 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.699762 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2xf4\" (UniqueName: \"kubernetes.io/projected/ee95ba9f-567a-461d-88c1-d3a904c366ad-kube-api-access-c2xf4\") pod \"tuned-wq59k\" (UID: \"ee95ba9f-567a-461d-88c1-d3a904c366ad\") " pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.700096 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.700075 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkrn6\" (UniqueName: \"kubernetes.io/projected/2b26247c-8561-4d36-a271-45bd1c30c698-kube-api-access-hkrn6\") pod \"node-ca-r48qq\" (UID: \"2b26247c-8561-4d36-a271-45bd1c30c698\") " pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:24.784305 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.784208 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:24.793175 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.793137 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" Apr 22 13:21:24.801113 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.800919 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-grkk2" Apr 22 13:21:24.807855 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.807828 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-hv7gn" Apr 22 13:21:24.814543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.814516 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" Apr 22 13:21:24.820260 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.820240 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:24.827915 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.827888 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-wq59k" Apr 22 13:21:24.832541 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:24.832518 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-r48qq" Apr 22 13:21:25.094969 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.094879 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:25.094969 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.094922 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:25.095199 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.095052 2579 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:25.095199 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.095063 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 13:21:25.095199 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.095086 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 13:21:25.095199 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.095099 2579 projected.go:194] Error preparing data for projected volume kube-api-access-zbhdt for pod openshift-network-diagnostics/network-check-target-tmp8r: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:25.095199 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.095116 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs podName:5dec8b65-6274-44b1-a169-ad20045acd1a nodeName:}" failed. No retries permitted until 2026-04-22 13:21:26.095097286 +0000 UTC m=+4.031484131 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs") pod "network-metrics-daemon-8vcg7" (UID: "5dec8b65-6274-44b1-a169-ad20045acd1a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:25.095199 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.095149 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt podName:4829e038-c6a2-4934-8182-fae81fb059a8 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:26.095133291 +0000 UTC m=+4.031520132 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-zbhdt" (UniqueName: "kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt") pod "network-check-target-tmp8r" (UID: "4829e038-c6a2-4934-8182-fae81fb059a8") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:25.195522 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.195484 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:25.195686 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.195655 2579 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:25.195741 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.195732 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret podName:6560c958-0928-4fa1-8691-8fe812c8d951 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:26.195709968 +0000 UTC m=+4.132096834 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret") pod "global-pull-secret-syncer-c7j5x" (UID: "6560c958-0928-4fa1-8691-8fe812c8d951") : object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:25.310482 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:25.310450 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea7df7e2_bf24_4c31_92ce_352960562242.slice/crio-abef903c45c048fe62e5423ebec283812e17f6ce35f887e6a38fb6190db729a8 WatchSource:0}: Error finding container abef903c45c048fe62e5423ebec283812e17f6ce35f887e6a38fb6190db729a8: Status 404 returned error can't find the container with id abef903c45c048fe62e5423ebec283812e17f6ce35f887e6a38fb6190db729a8 Apr 22 13:21:25.311456 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:25.311418 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4457d5d_3460_41bb_a335_8f9d5e2a4197.slice/crio-4b001d1c77a2f01328e8e3854c0908470c526289959504477e8e025c99b890a8 WatchSource:0}: Error finding container 4b001d1c77a2f01328e8e3854c0908470c526289959504477e8e025c99b890a8: Status 404 returned error can't find the container with id 4b001d1c77a2f01328e8e3854c0908470c526289959504477e8e025c99b890a8 Apr 22 13:21:25.312523 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:25.312496 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87ad5783_96fb_40e6_abe4_0fda9e9532b6.slice/crio-654fa5ae36f56388d188059857c0deb85c015c7d942f73cbab9c08e5835d0297 WatchSource:0}: Error finding container 654fa5ae36f56388d188059857c0deb85c015c7d942f73cbab9c08e5835d0297: Status 404 returned error can't find the container with id 654fa5ae36f56388d188059857c0deb85c015c7d942f73cbab9c08e5835d0297 Apr 22 13:21:25.316386 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:25.316361 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b26247c_8561_4d36_a271_45bd1c30c698.slice/crio-8b9f2215caf594b09c55ec893181a4338abd1a2e6871f4fda6dbe615effcfbb9 WatchSource:0}: Error finding container 8b9f2215caf594b09c55ec893181a4338abd1a2e6871f4fda6dbe615effcfbb9: Status 404 returned error can't find the container with id 8b9f2215caf594b09c55ec893181a4338abd1a2e6871f4fda6dbe615effcfbb9 Apr 22 13:21:25.317353 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:25.317334 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee95ba9f_567a_461d_88c1_d3a904c366ad.slice/crio-ea8ee131df9e5b013da9f1a79941435fb36056dd7b40dd3b54d885bfb0a6d42f WatchSource:0}: Error finding container ea8ee131df9e5b013da9f1a79941435fb36056dd7b40dd3b54d885bfb0a6d42f: Status 404 returned error can't find the container with id ea8ee131df9e5b013da9f1a79941435fb36056dd7b40dd3b54d885bfb0a6d42f Apr 22 13:21:25.318486 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:25.318462 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d77049f_f54b_436c_8e41_5e0b8d54e8bf.slice/crio-965376ebcf49b007193561d6bea443a17731a7173ce817adf9c35a9855d53e28 WatchSource:0}: Error finding container 965376ebcf49b007193561d6bea443a17731a7173ce817adf9c35a9855d53e28: Status 404 returned error can't find the container with id 965376ebcf49b007193561d6bea443a17731a7173ce817adf9c35a9855d53e28 Apr 22 13:21:25.514641 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.514425 2579 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 13:16:23 +0000 UTC" deadline="2027-12-13 18:50:20.400875145 +0000 UTC" Apr 22 13:21:25.514641 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.514636 2579 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14405h28m54.886243281s" Apr 22 13:21:25.601158 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.601047 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:25.601158 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:25.601151 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:25.608347 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.608309 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"0617d6f116f4c29dd9b84a31e919e0c1243e0a04731d2bc395828094dbd0d0a0"} Apr 22 13:21:25.609449 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.609412 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-hv7gn" event={"ID":"f757ef9d-b853-489b-9d90-79295660b525","Type":"ContainerStarted","Data":"3e2e6d1fe94566eafe2f53438e5c0e21381d71d9b4e69cdae97a7dd92722b987"} Apr 22 13:21:25.610521 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.610501 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-r48qq" event={"ID":"2b26247c-8561-4d36-a271-45bd1c30c698","Type":"ContainerStarted","Data":"8b9f2215caf594b09c55ec893181a4338abd1a2e6871f4fda6dbe615effcfbb9"} Apr 22 13:21:25.611471 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.611418 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-hrtnk" event={"ID":"87ad5783-96fb-40e6-abe4-0fda9e9532b6","Type":"ContainerStarted","Data":"654fa5ae36f56388d188059857c0deb85c015c7d942f73cbab9c08e5835d0297"} Apr 22 13:21:25.612389 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.612372 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" event={"ID":"ea7df7e2-bf24-4c31-92ce-352960562242","Type":"ContainerStarted","Data":"abef903c45c048fe62e5423ebec283812e17f6ce35f887e6a38fb6190db729a8"} Apr 22 13:21:25.614195 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.614174 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" event={"ID":"16c68cd8f43f9f069938f5a899c05cc7","Type":"ContainerStarted","Data":"621367e2af5bd952d3f71ceb6dac62cd6c55acc5557b7d3dbb8c0dd9f5e0c2e9"} Apr 22 13:21:25.615302 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.615280 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerStarted","Data":"965376ebcf49b007193561d6bea443a17731a7173ce817adf9c35a9855d53e28"} Apr 22 13:21:25.616379 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.616362 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-wq59k" event={"ID":"ee95ba9f-567a-461d-88c1-d3a904c366ad","Type":"ContainerStarted","Data":"ea8ee131df9e5b013da9f1a79941435fb36056dd7b40dd3b54d885bfb0a6d42f"} Apr 22 13:21:25.617326 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.617306 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-grkk2" event={"ID":"a4457d5d-3460-41bb-a335-8f9d5e2a4197","Type":"ContainerStarted","Data":"4b001d1c77a2f01328e8e3854c0908470c526289959504477e8e025c99b890a8"} Apr 22 13:21:25.626449 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:25.626385 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-128-54.ec2.internal" podStartSLOduration=2.626373258 podStartE2EDuration="2.626373258s" podCreationTimestamp="2026-04-22 13:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 13:21:25.625788492 +0000 UTC m=+3.562175354" watchObservedRunningTime="2026-04-22 13:21:25.626373258 +0000 UTC m=+3.562760121" Apr 22 13:21:26.103809 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:26.103773 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:26.103963 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:26.103830 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:26.104019 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.103964 2579 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:26.104073 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.104023 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs podName:5dec8b65-6274-44b1-a169-ad20045acd1a nodeName:}" failed. No retries permitted until 2026-04-22 13:21:28.104005527 +0000 UTC m=+6.040392370 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs") pod "network-metrics-daemon-8vcg7" (UID: "5dec8b65-6274-44b1-a169-ad20045acd1a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:26.104484 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.104467 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 13:21:26.104585 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.104491 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 13:21:26.104585 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.104504 2579 projected.go:194] Error preparing data for projected volume kube-api-access-zbhdt for pod openshift-network-diagnostics/network-check-target-tmp8r: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:26.104585 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.104550 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt podName:4829e038-c6a2-4934-8182-fae81fb059a8 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:28.104534024 +0000 UTC m=+6.040920872 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-zbhdt" (UniqueName: "kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt") pod "network-check-target-tmp8r" (UID: "4829e038-c6a2-4934-8182-fae81fb059a8") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:26.205533 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:26.204248 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:26.205533 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.204394 2579 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:26.205533 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.204503 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret podName:6560c958-0928-4fa1-8691-8fe812c8d951 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:28.204452994 +0000 UTC m=+6.140839840 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret") pod "global-pull-secret-syncer-c7j5x" (UID: "6560c958-0928-4fa1-8691-8fe812c8d951") : object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:26.600877 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:26.600791 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:26.601365 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.600933 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:26.601424 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:26.601405 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:26.601575 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:26.601532 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:26.624979 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:26.624611 2579 generic.go:358] "Generic (PLEG): container finished" podID="567021ad3684c4b7c9c2e8061f7ecdc6" containerID="8e05e2ade037fdf35609f82ca22919cd9030d865725b44334f1ff5d26447927d" exitCode=0 Apr 22 13:21:26.624979 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:26.624752 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" event={"ID":"567021ad3684c4b7c9c2e8061f7ecdc6","Type":"ContainerDied","Data":"8e05e2ade037fdf35609f82ca22919cd9030d865725b44334f1ff5d26447927d"} Apr 22 13:21:27.532757 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.532721 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5vr8r"] Apr 22 13:21:27.535942 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.535906 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.539092 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.538263 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-bwhcj\"" Apr 22 13:21:27.539092 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.538708 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 22 13:21:27.540943 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.539396 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 22 13:21:27.600387 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.600352 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:27.600581 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:27.600502 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:27.617230 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.617181 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z25ph\" (UniqueName: \"kubernetes.io/projected/c55e4ce3-f705-4675-8b13-55cf25d10b37-kube-api-access-z25ph\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.617692 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.617247 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c55e4ce3-f705-4675-8b13-55cf25d10b37-hosts-file\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.617692 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.617298 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c55e4ce3-f705-4675-8b13-55cf25d10b37-tmp-dir\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.631296 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.631255 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" event={"ID":"567021ad3684c4b7c9c2e8061f7ecdc6","Type":"ContainerStarted","Data":"08ad0ae3fc689a026131272716f9ce049c2cb5fc049edfa0fa1cdc2cf4b57282"} Apr 22 13:21:27.645118 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.645068 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-128-54.ec2.internal" podStartSLOduration=4.6450487240000005 podStartE2EDuration="4.645048724s" podCreationTimestamp="2026-04-22 13:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 13:21:27.64409574 +0000 UTC m=+5.580482627" watchObservedRunningTime="2026-04-22 13:21:27.645048724 +0000 UTC m=+5.581435587" Apr 22 13:21:27.717808 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.717771 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-z25ph\" (UniqueName: \"kubernetes.io/projected/c55e4ce3-f705-4675-8b13-55cf25d10b37-kube-api-access-z25ph\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.717993 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.717828 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c55e4ce3-f705-4675-8b13-55cf25d10b37-hosts-file\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.717993 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.717877 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c55e4ce3-f705-4675-8b13-55cf25d10b37-tmp-dir\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.718214 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.718187 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c55e4ce3-f705-4675-8b13-55cf25d10b37-hosts-file\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.718295 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.718273 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c55e4ce3-f705-4675-8b13-55cf25d10b37-tmp-dir\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.734641 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.734612 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-z25ph\" (UniqueName: \"kubernetes.io/projected/c55e4ce3-f705-4675-8b13-55cf25d10b37-kube-api-access-z25ph\") pod \"node-resolver-5vr8r\" (UID: \"c55e4ce3-f705-4675-8b13-55cf25d10b37\") " pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:27.848712 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:27.848625 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5vr8r" Apr 22 13:21:28.122479 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:28.122369 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:28.122479 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:28.122426 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:28.122756 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.122572 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 13:21:28.122756 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.122598 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 13:21:28.122756 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.122612 2579 projected.go:194] Error preparing data for projected volume kube-api-access-zbhdt for pod openshift-network-diagnostics/network-check-target-tmp8r: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:28.122756 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.122617 2579 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:28.122756 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.122673 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt podName:4829e038-c6a2-4934-8182-fae81fb059a8 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:32.12265463 +0000 UTC m=+10.059041481 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-zbhdt" (UniqueName: "kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt") pod "network-check-target-tmp8r" (UID: "4829e038-c6a2-4934-8182-fae81fb059a8") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:28.122756 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.122692 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs podName:5dec8b65-6274-44b1-a169-ad20045acd1a nodeName:}" failed. No retries permitted until 2026-04-22 13:21:32.122683228 +0000 UTC m=+10.059070075 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs") pod "network-metrics-daemon-8vcg7" (UID: "5dec8b65-6274-44b1-a169-ad20045acd1a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:28.222938 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:28.222899 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:28.223103 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.223092 2579 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:28.223164 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.223156 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret podName:6560c958-0928-4fa1-8691-8fe812c8d951 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:32.223137479 +0000 UTC m=+10.159524325 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret") pod "global-pull-secret-syncer-c7j5x" (UID: "6560c958-0928-4fa1-8691-8fe812c8d951") : object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:28.605356 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:28.605271 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:28.605532 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.605403 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:28.605868 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:28.605842 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:28.605961 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:28.605943 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:29.601081 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:29.601035 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:29.601631 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:29.601188 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:30.603857 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:30.603828 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:30.603857 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:30.603847 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:30.604368 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:30.603966 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:30.604461 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:30.604394 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:31.600361 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:31.600323 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:31.600578 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:31.600484 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:32.156739 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:32.156688 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:32.156739 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:32.156737 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:32.157274 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.156849 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 13:21:32.157274 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.156871 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 13:21:32.157274 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.156886 2579 projected.go:194] Error preparing data for projected volume kube-api-access-zbhdt for pod openshift-network-diagnostics/network-check-target-tmp8r: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:32.157274 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.156919 2579 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:32.157274 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.156940 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt podName:4829e038-c6a2-4934-8182-fae81fb059a8 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:40.156924174 +0000 UTC m=+18.093311020 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-zbhdt" (UniqueName: "kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt") pod "network-check-target-tmp8r" (UID: "4829e038-c6a2-4934-8182-fae81fb059a8") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:32.157274 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.156978 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs podName:5dec8b65-6274-44b1-a169-ad20045acd1a nodeName:}" failed. No retries permitted until 2026-04-22 13:21:40.15696073 +0000 UTC m=+18.093347588 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs") pod "network-metrics-daemon-8vcg7" (UID: "5dec8b65-6274-44b1-a169-ad20045acd1a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:32.258828 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:32.258265 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:32.258828 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.258453 2579 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:32.258828 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.258511 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret podName:6560c958-0928-4fa1-8691-8fe812c8d951 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:40.258495087 +0000 UTC m=+18.194881941 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret") pod "global-pull-secret-syncer-c7j5x" (UID: "6560c958-0928-4fa1-8691-8fe812c8d951") : object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:32.604395 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:32.604366 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:32.604395 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:32.604390 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:32.604637 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.604500 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:32.604637 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:32.604623 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:33.600688 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:33.600654 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:33.601140 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:33.600794 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:34.600734 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:34.600644 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:34.601154 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:34.600782 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:34.601154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:34.600796 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:34.601154 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:34.600894 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:35.600779 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:35.600741 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:35.601217 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:35.600870 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:36.603575 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:36.603542 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:36.604054 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:36.603542 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:36.604054 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:36.603690 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:36.604054 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:36.603727 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:37.600369 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:37.600327 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:37.600677 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:37.600488 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:38.600414 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:38.600381 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:38.600812 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:38.600381 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:38.600812 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:38.600533 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:38.600812 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:38.600571 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:39.600877 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:39.600841 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:39.601327 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:39.600959 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:40.213684 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:40.213647 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:40.213905 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:40.213698 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:40.213905 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.213810 2579 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:40.213905 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.213827 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 13:21:40.213905 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.213855 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 13:21:40.213905 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.213868 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs podName:5dec8b65-6274-44b1-a169-ad20045acd1a nodeName:}" failed. No retries permitted until 2026-04-22 13:21:56.213851646 +0000 UTC m=+34.150238491 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs") pod "network-metrics-daemon-8vcg7" (UID: "5dec8b65-6274-44b1-a169-ad20045acd1a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:40.213905 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.213868 2579 projected.go:194] Error preparing data for projected volume kube-api-access-zbhdt for pod openshift-network-diagnostics/network-check-target-tmp8r: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:40.213905 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.213899 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt podName:4829e038-c6a2-4934-8182-fae81fb059a8 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:56.213892935 +0000 UTC m=+34.150279776 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-zbhdt" (UniqueName: "kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt") pod "network-check-target-tmp8r" (UID: "4829e038-c6a2-4934-8182-fae81fb059a8") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:40.315026 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:40.314989 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:40.315215 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.315114 2579 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:40.315215 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.315179 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret podName:6560c958-0928-4fa1-8691-8fe812c8d951 nodeName:}" failed. No retries permitted until 2026-04-22 13:21:56.315161466 +0000 UTC m=+34.251548324 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret") pod "global-pull-secret-syncer-c7j5x" (UID: "6560c958-0928-4fa1-8691-8fe812c8d951") : object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:40.600668 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:40.600579 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:40.600839 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.600690 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:40.600839 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:40.600777 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:40.601231 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:40.600888 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:41.600141 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:41.600106 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:41.600305 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:41.600210 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:41.760038 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:21:41.759991 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc55e4ce3_f705_4675_8b13_55cf25d10b37.slice/crio-9d2411152801ccab86c5bc2aa38e5108187b3a27ffb3b95144b7e2d96d6bb89d WatchSource:0}: Error finding container 9d2411152801ccab86c5bc2aa38e5108187b3a27ffb3b95144b7e2d96d6bb89d: Status 404 returned error can't find the container with id 9d2411152801ccab86c5bc2aa38e5108187b3a27ffb3b95144b7e2d96d6bb89d Apr 22 13:21:42.601144 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.600966 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:42.601300 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.601032 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:42.601361 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:42.601201 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:42.601361 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:42.601332 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:42.659001 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.658894 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-r48qq" event={"ID":"2b26247c-8561-4d36-a271-45bd1c30c698","Type":"ContainerStarted","Data":"e6b87c678964ba255ac79efd2210c122f3dcc09942dbffe5344cbedfae303396"} Apr 22 13:21:42.660206 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.660180 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-hrtnk" event={"ID":"87ad5783-96fb-40e6-abe4-0fda9e9532b6","Type":"ContainerStarted","Data":"9a4ef7404a2445725040a0922b39837ee66f9955f09ed8ccdd2c0522b309d677"} Apr 22 13:21:42.661351 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.661327 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" event={"ID":"ea7df7e2-bf24-4c31-92ce-352960562242","Type":"ContainerStarted","Data":"08da258f2c6ec4cbd68c5b37cb22cc0f3d76ca1d748d402200e4fc9e1e76a292"} Apr 22 13:21:42.662556 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.662533 2579 generic.go:358] "Generic (PLEG): container finished" podID="1d77049f-f54b-436c-8e41-5e0b8d54e8bf" containerID="2b0cb53c2b750fde92cef95bf2e3ce5b3e1e7e26958ef7b52a950d6669ac24b9" exitCode=0 Apr 22 13:21:42.662659 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.662618 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerDied","Data":"2b0cb53c2b750fde92cef95bf2e3ce5b3e1e7e26958ef7b52a950d6669ac24b9"} Apr 22 13:21:42.663892 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.663868 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-wq59k" event={"ID":"ee95ba9f-567a-461d-88c1-d3a904c366ad","Type":"ContainerStarted","Data":"bcc15b3130049853e1ae768acdb190dd6198f9e4d259a2e4031d08edf187a01e"} Apr 22 13:21:42.665147 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.665129 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-grkk2" event={"ID":"a4457d5d-3460-41bb-a335-8f9d5e2a4197","Type":"ContainerStarted","Data":"c4e8c662424fd63b658f5609c1451bcc5bf2d1bba88bce8ea04bbc65fc2604c9"} Apr 22 13:21:42.666355 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.666337 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5vr8r" event={"ID":"c55e4ce3-f705-4675-8b13-55cf25d10b37","Type":"ContainerStarted","Data":"bdffe99dc6c878c28a9ab8994afe07121cc0d8661ccc6a9a60ec279377d9d174"} Apr 22 13:21:42.666460 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.666360 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5vr8r" event={"ID":"c55e4ce3-f705-4675-8b13-55cf25d10b37","Type":"ContainerStarted","Data":"9d2411152801ccab86c5bc2aa38e5108187b3a27ffb3b95144b7e2d96d6bb89d"} Apr 22 13:21:42.668173 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.668157 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:21:42.668393 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.668376 2579 generic.go:358] "Generic (PLEG): container finished" podID="782146a0-22aa-427e-aeb7-43bb6980c291" containerID="31ac4c97486ba518a4afc4ab1758ddfc8ed6ddc43caefd310a3da6dc1e2a1854" exitCode=1 Apr 22 13:21:42.668466 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.668404 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"c5d6413852370b6da71359b7303e9eab51f648ebec1419465394e7e6b28f76be"} Apr 22 13:21:42.668466 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.668417 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"befe4e3caab010f35ec88007c732a5304ed6299ae570615f6e093a556f506484"} Apr 22 13:21:42.668466 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.668440 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"62c0276aac97395034b46d5e0ac9919b80730814becd5bf67732166d463742ee"} Apr 22 13:21:42.668466 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.668454 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerDied","Data":"31ac4c97486ba518a4afc4ab1758ddfc8ed6ddc43caefd310a3da6dc1e2a1854"} Apr 22 13:21:42.668606 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.668467 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"f29c1e10ed8fd1223cba9443e31abac79feab2dadc41312854616e694c9c846e"} Apr 22 13:21:42.673584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.673553 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-r48qq" podStartSLOduration=8.529528754 podStartE2EDuration="20.673541507s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.318389182 +0000 UTC m=+3.254776025" lastFinishedPulling="2026-04-22 13:21:37.46240192 +0000 UTC m=+15.398788778" observedRunningTime="2026-04-22 13:21:42.673141487 +0000 UTC m=+20.609528361" watchObservedRunningTime="2026-04-22 13:21:42.673541507 +0000 UTC m=+20.609928369" Apr 22 13:21:42.691011 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.690969 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-wq59k" podStartSLOduration=4.233815091 podStartE2EDuration="20.690955595s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.319818248 +0000 UTC m=+3.256205097" lastFinishedPulling="2026-04-22 13:21:41.776958755 +0000 UTC m=+19.713345601" observedRunningTime="2026-04-22 13:21:42.690384671 +0000 UTC m=+20.626771534" watchObservedRunningTime="2026-04-22 13:21:42.690955595 +0000 UTC m=+20.627342458" Apr 22 13:21:42.705476 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.705441 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5vr8r" podStartSLOduration=15.705417409 podStartE2EDuration="15.705417409s" podCreationTimestamp="2026-04-22 13:21:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 13:21:42.705339702 +0000 UTC m=+20.641726566" watchObservedRunningTime="2026-04-22 13:21:42.705417409 +0000 UTC m=+20.641804271" Apr 22 13:21:42.726050 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.725782 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-grkk2" podStartSLOduration=4.175102809 podStartE2EDuration="20.725764147s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.314622899 +0000 UTC m=+3.251009755" lastFinishedPulling="2026-04-22 13:21:41.865284236 +0000 UTC m=+19.801671093" observedRunningTime="2026-04-22 13:21:42.725656747 +0000 UTC m=+20.662043613" watchObservedRunningTime="2026-04-22 13:21:42.725764147 +0000 UTC m=+20.662151013" Apr 22 13:21:42.761717 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:42.761658 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-hrtnk" podStartSLOduration=4.313389387 podStartE2EDuration="20.761634855s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.314747824 +0000 UTC m=+3.251134668" lastFinishedPulling="2026-04-22 13:21:41.762993292 +0000 UTC m=+19.699380136" observedRunningTime="2026-04-22 13:21:42.761173463 +0000 UTC m=+20.697560326" watchObservedRunningTime="2026-04-22 13:21:42.761634855 +0000 UTC m=+20.698021720" Apr 22 13:21:43.600543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.600516 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:43.600728 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:43.600632 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:43.667560 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.667531 2579 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 22 13:21:43.673900 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.673876 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:21:43.674254 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.674225 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"8825af042279536d1bcaa1251142d11827192f2087442ae070d65d20636a9d32"} Apr 22 13:21:43.675647 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.675589 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-hv7gn" event={"ID":"f757ef9d-b853-489b-9d90-79295660b525","Type":"ContainerStarted","Data":"2b295511664833d7f1c541f7589a49e53073dc4089e42cfbca40b794162711db"} Apr 22 13:21:43.677891 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.677865 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" event={"ID":"ea7df7e2-bf24-4c31-92ce-352960562242","Type":"ContainerStarted","Data":"d078b98c20762bd08d351bed9f76b81b7414041a5d6608d3f5459e12d1d2970a"} Apr 22 13:21:43.691176 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.691120 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-hv7gn" podStartSLOduration=5.245968495 podStartE2EDuration="21.691100137s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.322824167 +0000 UTC m=+3.259211012" lastFinishedPulling="2026-04-22 13:21:41.767955799 +0000 UTC m=+19.704342654" observedRunningTime="2026-04-22 13:21:43.690981943 +0000 UTC m=+21.627368806" watchObservedRunningTime="2026-04-22 13:21:43.691100137 +0000 UTC m=+21.627487003" Apr 22 13:21:43.937781 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:43.937711 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-5vr8r_c55e4ce3-f705-4675-8b13-55cf25d10b37/dns-node-resolver/0.log" Apr 22 13:21:44.531945 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:44.531821 2579 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-22T13:21:43.667553023Z","UUID":"59747353-8368-4fd6-8a2c-46962e7492c8","Handler":null,"Name":"","Endpoint":""} Apr 22 13:21:44.533797 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:44.533747 2579 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 22 13:21:44.533946 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:44.533806 2579 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 22 13:21:44.600470 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:44.600425 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:44.600643 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:44.600577 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:44.600715 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:44.600635 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:44.600787 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:44.600760 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:45.028664 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.028339 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:45.029094 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.029053 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:45.307686 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.307592 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:45.308091 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.308073 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-hrtnk" Apr 22 13:21:45.316697 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.316673 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-r48qq_2b26247c-8561-4d36-a271-45bd1c30c698/node-ca/0.log" Apr 22 13:21:45.600144 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.600058 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:45.600320 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:45.600179 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:45.683333 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.683302 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" event={"ID":"ea7df7e2-bf24-4c31-92ce-352960562242","Type":"ContainerStarted","Data":"103983480eb1fdc38253a8d1d24247848600527d33177a63b382a57992213a4e"} Apr 22 13:21:45.686204 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.686178 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:21:45.686552 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:45.686530 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"279126ecb447ea46724034517d90845f62a8e65e4a64d667a8ed924ad8525615"} Apr 22 13:21:46.601047 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:46.601012 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:46.601697 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:46.601027 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:46.601697 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:46.601145 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:46.601697 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:46.601222 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:47.600954 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.600770 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:47.601125 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:47.601022 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:47.692888 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.692860 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:21:47.693242 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.693210 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"eacbaec0f3bcc1e5e807b75ba9725e9d93ec72927f20de4397e89c78e48fa698"} Apr 22 13:21:47.693534 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.693513 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:47.693643 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.693544 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:47.693830 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.693814 2579 scope.go:117] "RemoveContainer" containerID="31ac4c97486ba518a4afc4ab1758ddfc8ed6ddc43caefd310a3da6dc1e2a1854" Apr 22 13:21:47.694947 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.694928 2579 generic.go:358] "Generic (PLEG): container finished" podID="1d77049f-f54b-436c-8e41-5e0b8d54e8bf" containerID="65287f5d67530c71919dbb4a046d0622126fa1288ffad45f3bfa998926f0dfbd" exitCode=0 Apr 22 13:21:47.695029 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.694963 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerDied","Data":"65287f5d67530c71919dbb4a046d0622126fa1288ffad45f3bfa998926f0dfbd"} Apr 22 13:21:47.710295 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.710268 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:47.725031 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:47.724983 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-qz6ss" podStartSLOduration=6.119535641 podStartE2EDuration="25.724970196s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.313121728 +0000 UTC m=+3.249508583" lastFinishedPulling="2026-04-22 13:21:44.918556295 +0000 UTC m=+22.854943138" observedRunningTime="2026-04-22 13:21:45.703665297 +0000 UTC m=+23.640052163" watchObservedRunningTime="2026-04-22 13:21:47.724970196 +0000 UTC m=+25.661357058" Apr 22 13:21:48.601060 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.601033 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:48.601145 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.601070 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:48.601407 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:48.601170 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:48.601407 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:48.601267 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:48.699465 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.699419 2579 generic.go:358] "Generic (PLEG): container finished" podID="1d77049f-f54b-436c-8e41-5e0b8d54e8bf" containerID="ad7dc774cdcf55ca29feae2dab0ac976dbf3f078a4063141135a6eb69f8595db" exitCode=0 Apr 22 13:21:48.699628 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.699501 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerDied","Data":"ad7dc774cdcf55ca29feae2dab0ac976dbf3f078a4063141135a6eb69f8595db"} Apr 22 13:21:48.702741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.702725 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:21:48.703072 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.703038 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" event={"ID":"782146a0-22aa-427e-aeb7-43bb6980c291","Type":"ContainerStarted","Data":"41ee0cbf56cd241711bce12317c5ef321036fe6be6fa2aaa7cfce9c83643e1f8"} Apr 22 13:21:48.703306 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.703289 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:48.717973 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.717895 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:21:48.749636 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:48.749593 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" podStartSLOduration=10.117921014 podStartE2EDuration="26.749580574s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.324023971 +0000 UTC m=+3.260410820" lastFinishedPulling="2026-04-22 13:21:41.955683533 +0000 UTC m=+19.892070380" observedRunningTime="2026-04-22 13:21:48.749288408 +0000 UTC m=+26.685675305" watchObservedRunningTime="2026-04-22 13:21:48.749580574 +0000 UTC m=+26.685967436" Apr 22 13:21:49.600684 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:49.600659 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:49.600808 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:49.600785 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:49.706601 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:49.706572 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerStarted","Data":"beb657db68ab24f931e43d44cda6735ae88626e6bf602b346a6cf697ae8e72f9"} Apr 22 13:21:50.600568 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:50.600538 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:50.600760 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:50.600638 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:50.600760 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:50.600698 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:50.600839 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:50.600792 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:50.709930 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:50.709896 2579 generic.go:358] "Generic (PLEG): container finished" podID="1d77049f-f54b-436c-8e41-5e0b8d54e8bf" containerID="beb657db68ab24f931e43d44cda6735ae88626e6bf602b346a6cf697ae8e72f9" exitCode=0 Apr 22 13:21:50.710387 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:50.709930 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerDied","Data":"beb657db68ab24f931e43d44cda6735ae88626e6bf602b346a6cf697ae8e72f9"} Apr 22 13:21:51.600729 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:51.600684 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:51.600937 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:51.600841 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:52.601372 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:52.601335 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:52.601785 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:52.601532 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:52.601785 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:52.601577 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:52.601785 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:52.601683 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:53.600068 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:53.600027 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:53.600253 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:53.600156 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:54.600659 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:54.600613 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:54.601193 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:54.600638 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:54.601193 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:54.600750 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:54.601193 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:54.600805 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:55.600483 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:55.600423 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:55.600695 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:55.600560 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:56.242650 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:56.242616 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:56.242650 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:56.242653 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:56.242936 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.242768 2579 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:56.242936 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.242790 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 13:21:56.242936 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.242808 2579 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 13:21:56.242936 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.242819 2579 projected.go:194] Error preparing data for projected volume kube-api-access-zbhdt for pod openshift-network-diagnostics/network-check-target-tmp8r: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:56.242936 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.242830 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs podName:5dec8b65-6274-44b1-a169-ad20045acd1a nodeName:}" failed. No retries permitted until 2026-04-22 13:22:28.242810667 +0000 UTC m=+66.179197509 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs") pod "network-metrics-daemon-8vcg7" (UID: "5dec8b65-6274-44b1-a169-ad20045acd1a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 13:21:56.242936 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.242853 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt podName:4829e038-c6a2-4934-8182-fae81fb059a8 nodeName:}" failed. No retries permitted until 2026-04-22 13:22:28.242843852 +0000 UTC m=+66.179230710 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-zbhdt" (UniqueName: "kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt") pod "network-check-target-tmp8r" (UID: "4829e038-c6a2-4934-8182-fae81fb059a8") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 13:21:56.343833 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:56.343803 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:56.343977 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.343914 2579 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:56.343977 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.343961 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret podName:6560c958-0928-4fa1-8691-8fe812c8d951 nodeName:}" failed. No retries permitted until 2026-04-22 13:22:28.343947738 +0000 UTC m=+66.280334579 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret") pod "global-pull-secret-syncer-c7j5x" (UID: "6560c958-0928-4fa1-8691-8fe812c8d951") : object "kube-system"/"original-pull-secret" not registered Apr 22 13:21:56.600921 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:56.600752 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:56.601219 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:56.600783 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:56.601219 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.601014 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:56.601219 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:56.601082 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:57.600682 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:57.600647 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:57.600866 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:57.600744 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:57.728442 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:57.728392 2579 generic.go:358] "Generic (PLEG): container finished" podID="1d77049f-f54b-436c-8e41-5e0b8d54e8bf" containerID="3ca05faf778ecd28d41c0223148d3f79e270f955e9cc5b2718663266111083d2" exitCode=0 Apr 22 13:21:57.728826 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:57.728475 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerDied","Data":"3ca05faf778ecd28d41c0223148d3f79e270f955e9cc5b2718663266111083d2"} Apr 22 13:21:58.601035 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:58.600999 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:21:58.601274 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:58.601070 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:21:58.601274 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:58.601191 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:21:58.601384 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:58.601305 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:21:58.732963 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:58.732929 2579 generic.go:358] "Generic (PLEG): container finished" podID="1d77049f-f54b-436c-8e41-5e0b8d54e8bf" containerID="65cc3c352e122fc00b97a67ea0f7cde1acfc6b8860ee163bb1c6a886a67cb6ab" exitCode=0 Apr 22 13:21:58.733308 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:58.732993 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerDied","Data":"65cc3c352e122fc00b97a67ea0f7cde1acfc6b8860ee163bb1c6a886a67cb6ab"} Apr 22 13:21:59.600990 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:59.600958 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:21:59.601164 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:21:59.601049 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:21:59.737269 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:59.737231 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" event={"ID":"1d77049f-f54b-436c-8e41-5e0b8d54e8bf","Type":"ContainerStarted","Data":"4264b792a0e40e50955bb9ed3bf366503d25064f91d12a307fdcbd7abfee4250"} Apr 22 13:21:59.759689 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:21:59.759647 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vzwxc" podStartSLOduration=6.514242231 podStartE2EDuration="37.759630213s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:21:25.322208248 +0000 UTC m=+3.258595103" lastFinishedPulling="2026-04-22 13:21:56.567596242 +0000 UTC m=+34.503983085" observedRunningTime="2026-04-22 13:21:59.758698583 +0000 UTC m=+37.695085446" watchObservedRunningTime="2026-04-22 13:21:59.759630213 +0000 UTC m=+37.696017054" Apr 22 13:22:00.600111 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:00.600084 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:00.600279 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:00.600129 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:00.600279 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:00.600218 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:22:00.600365 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:00.600322 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:22:01.600684 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:01.600637 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:01.601079 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:01.600751 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:22:02.601322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:02.601294 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:02.601797 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:02.601372 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:22:02.601797 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:02.601484 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:02.601797 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:02.601604 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:22:03.600621 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:03.600581 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:03.600807 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:03.600718 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:22:04.600404 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:04.600373 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:04.600929 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:04.600410 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:04.600929 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:04.600523 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:22:04.600929 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:04.600627 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:22:05.049258 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:05.049229 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-c7j5x"] Apr 22 13:22:05.049418 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:05.049327 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:05.049493 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:05.049411 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:22:05.055867 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:05.055813 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-tmp8r"] Apr 22 13:22:05.056064 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:05.055996 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:05.056159 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:05.056130 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:22:05.056387 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:05.056357 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8vcg7"] Apr 22 13:22:05.056574 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:05.056558 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:05.056805 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:05.056779 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:22:06.600009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:06.599974 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:06.600009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:06.599994 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:06.600633 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:06.599997 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:06.600633 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:06.600067 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:22:06.600633 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:06.600183 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:22:06.600633 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:06.600221 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:22:08.600536 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:08.600352 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:08.600953 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:08.600410 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:08.600953 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:08.600642 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-tmp8r" podUID="4829e038-c6a2-4934-8182-fae81fb059a8" Apr 22 13:22:08.600953 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:08.600493 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:08.600953 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:08.600678 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-c7j5x" podUID="6560c958-0928-4fa1-8691-8fe812c8d951" Apr 22 13:22:08.600953 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:08.600765 2579 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8vcg7" podUID="5dec8b65-6274-44b1-a169-ad20045acd1a" Apr 22 13:22:08.921808 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:08.921733 2579 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-128-54.ec2.internal" event="NodeReady" Apr 22 13:22:08.922005 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:08.921854 2579 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 22 13:22:08.962017 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:08.960755 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-756c9666d4-q6jfg"] Apr 22 13:22:09.002525 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.002494 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-q8tfx"] Apr 22 13:22:09.002695 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.002659 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.004879 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.004853 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 22 13:22:09.005009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.004883 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 22 13:22:09.005009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.004932 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-7w6d8\"" Apr 22 13:22:09.005144 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.005132 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 22 13:22:09.009360 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.009339 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 22 13:22:09.026326 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.026294 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-756c9666d4-q6jfg"] Apr 22 13:22:09.026326 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.026333 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-xg89z"] Apr 22 13:22:09.026512 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.026362 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.028959 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.028918 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-rrhbx\"" Apr 22 13:22:09.029096 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.029079 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 22 13:22:09.029159 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.029096 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 22 13:22:09.042401 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042385 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-registry-tls\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.042490 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042409 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-registry-certificates\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.042490 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042443 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-ca-trust-extracted\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.042490 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042460 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/82906f03-65d1-460f-b410-9f8526350920-metrics-tls\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.042590 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042505 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6chnv\" (UniqueName: \"kubernetes.io/projected/82906f03-65d1-460f-b410-9f8526350920-kube-api-access-6chnv\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.042590 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042534 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld86n\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-kube-api-access-ld86n\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.042590 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042552 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-trusted-ca\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.042590 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042582 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82906f03-65d1-460f-b410-9f8526350920-config-volume\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.042708 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042622 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/82906f03-65d1-460f-b410-9f8526350920-tmp-dir\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.042708 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042638 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-installation-pull-secrets\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.042708 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042653 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-bound-sa-token\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.042795 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.042713 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-image-registry-private-configuration\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.047846 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.047828 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-q8tfx"] Apr 22 13:22:09.047846 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.047849 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xg89z"] Apr 22 13:22:09.047953 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.047860 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-smwrp"] Apr 22 13:22:09.047989 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.047952 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.050463 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.050398 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 22 13:22:09.050463 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.050422 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-6bcgp\"" Apr 22 13:22:09.050463 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.050456 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 22 13:22:09.050755 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.050528 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 22 13:22:09.069747 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.069719 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-smwrp"] Apr 22 13:22:09.069862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.069755 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.072060 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.072041 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-jvn9v\"" Apr 22 13:22:09.072060 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.072054 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 22 13:22:09.072204 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.072077 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 22 13:22:09.072204 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.072149 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 22 13:22:09.072462 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.072447 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 22 13:22:09.143985 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.143949 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/512808ab-fbb1-46eb-86e9-a0fb60d532db-crio-socket\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.143985 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.143982 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/512808ab-fbb1-46eb-86e9-a0fb60d532db-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.144205 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144010 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnf96\" (UniqueName: \"kubernetes.io/projected/52d135c0-aa2e-4533-986a-5002ff68781d-kube-api-access-pnf96\") pod \"ingress-canary-xg89z\" (UID: \"52d135c0-aa2e-4533-986a-5002ff68781d\") " pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.144205 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144066 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/82906f03-65d1-460f-b410-9f8526350920-tmp-dir\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.144205 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144095 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-installation-pull-secrets\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144205 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144112 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-bound-sa-token\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144258 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-image-registry-private-configuration\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144292 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-registry-tls\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144320 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzt8z\" (UniqueName: \"kubernetes.io/projected/512808ab-fbb1-46eb-86e9-a0fb60d532db-kube-api-access-zzt8z\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.144515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144351 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-registry-certificates\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144380 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-ca-trust-extracted\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144383 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/82906f03-65d1-460f-b410-9f8526350920-tmp-dir\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.144515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144407 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/82906f03-65d1-460f-b410-9f8526350920-metrics-tls\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.144515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144450 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6chnv\" (UniqueName: \"kubernetes.io/projected/82906f03-65d1-460f-b410-9f8526350920-kube-api-access-6chnv\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.144515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144479 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ld86n\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-kube-api-access-ld86n\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144503 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-trusted-ca\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.144904 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144525 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/512808ab-fbb1-46eb-86e9-a0fb60d532db-data-volume\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.144904 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144547 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/512808ab-fbb1-46eb-86e9-a0fb60d532db-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.144904 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144579 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/52d135c0-aa2e-4533-986a-5002ff68781d-cert\") pod \"ingress-canary-xg89z\" (UID: \"52d135c0-aa2e-4533-986a-5002ff68781d\") " pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.144904 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144606 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82906f03-65d1-460f-b410-9f8526350920-config-volume\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.144904 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.144746 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-ca-trust-extracted\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.145186 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.145158 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82906f03-65d1-460f-b410-9f8526350920-config-volume\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.145337 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.145296 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-registry-certificates\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.145701 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.145679 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-trusted-ca\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.148487 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.148462 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/82906f03-65d1-460f-b410-9f8526350920-metrics-tls\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.148594 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.148496 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-registry-tls\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.148594 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.148577 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-installation-pull-secrets\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.148677 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.148634 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-image-registry-private-configuration\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.151973 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.151947 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-bound-sa-token\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.152198 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.152178 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6chnv\" (UniqueName: \"kubernetes.io/projected/82906f03-65d1-460f-b410-9f8526350920-kube-api-access-6chnv\") pod \"dns-default-q8tfx\" (UID: \"82906f03-65d1-460f-b410-9f8526350920\") " pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.152259 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.152207 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld86n\" (UniqueName: \"kubernetes.io/projected/d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f-kube-api-access-ld86n\") pod \"image-registry-756c9666d4-q6jfg\" (UID: \"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f\") " pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.245419 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245336 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/512808ab-fbb1-46eb-86e9-a0fb60d532db-data-volume\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.245419 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245377 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/512808ab-fbb1-46eb-86e9-a0fb60d532db-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.245419 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245402 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/52d135c0-aa2e-4533-986a-5002ff68781d-cert\") pod \"ingress-canary-xg89z\" (UID: \"52d135c0-aa2e-4533-986a-5002ff68781d\") " pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.245717 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245460 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/512808ab-fbb1-46eb-86e9-a0fb60d532db-crio-socket\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.245717 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245482 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/512808ab-fbb1-46eb-86e9-a0fb60d532db-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.245717 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245511 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pnf96\" (UniqueName: \"kubernetes.io/projected/52d135c0-aa2e-4533-986a-5002ff68781d-kube-api-access-pnf96\") pod \"ingress-canary-xg89z\" (UID: \"52d135c0-aa2e-4533-986a-5002ff68781d\") " pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.245717 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245576 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zzt8z\" (UniqueName: \"kubernetes.io/projected/512808ab-fbb1-46eb-86e9-a0fb60d532db-kube-api-access-zzt8z\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.245717 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.245670 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/512808ab-fbb1-46eb-86e9-a0fb60d532db-crio-socket\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.248017 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.247995 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/52d135c0-aa2e-4533-986a-5002ff68781d-cert\") pod \"ingress-canary-xg89z\" (UID: \"52d135c0-aa2e-4533-986a-5002ff68781d\") " pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.258152 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.258131 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/512808ab-fbb1-46eb-86e9-a0fb60d532db-data-volume\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.258353 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.258332 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/512808ab-fbb1-46eb-86e9-a0fb60d532db-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.258407 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.258356 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/512808ab-fbb1-46eb-86e9-a0fb60d532db-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.264790 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.264759 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzt8z\" (UniqueName: \"kubernetes.io/projected/512808ab-fbb1-46eb-86e9-a0fb60d532db-kube-api-access-zzt8z\") pod \"insights-runtime-extractor-smwrp\" (UID: \"512808ab-fbb1-46eb-86e9-a0fb60d532db\") " pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.264913 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.264764 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnf96\" (UniqueName: \"kubernetes.io/projected/52d135c0-aa2e-4533-986a-5002ff68781d-kube-api-access-pnf96\") pod \"ingress-canary-xg89z\" (UID: \"52d135c0-aa2e-4533-986a-5002ff68781d\") " pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.312256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.312218 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.334737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.334706 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:09.356366 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.356334 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xg89z" Apr 22 13:22:09.378150 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.378122 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-smwrp" Apr 22 13:22:09.506696 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.506664 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-756c9666d4-q6jfg"] Apr 22 13:22:09.509951 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.509930 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-q8tfx"] Apr 22 13:22:09.511704 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:09.511674 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd02cc9bc_fe6e_44ad_a130_545bbcd2fa0f.slice/crio-e7ffc9e95ff0b34b8001907ee5a73cf0d9613f945f25e09a8389bc9ea7d94ef8 WatchSource:0}: Error finding container e7ffc9e95ff0b34b8001907ee5a73cf0d9613f945f25e09a8389bc9ea7d94ef8: Status 404 returned error can't find the container with id e7ffc9e95ff0b34b8001907ee5a73cf0d9613f945f25e09a8389bc9ea7d94ef8 Apr 22 13:22:09.512783 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:09.512759 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82906f03_65d1_460f_b410_9f8526350920.slice/crio-e0035c031ffcf0c9c07e5797a5688f20a342d87a7826fdeca639eea2b03b7e2d WatchSource:0}: Error finding container e0035c031ffcf0c9c07e5797a5688f20a342d87a7826fdeca639eea2b03b7e2d: Status 404 returned error can't find the container with id e0035c031ffcf0c9c07e5797a5688f20a342d87a7826fdeca639eea2b03b7e2d Apr 22 13:22:09.524322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.524297 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xg89z"] Apr 22 13:22:09.527454 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:09.527416 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52d135c0_aa2e_4533_986a_5002ff68781d.slice/crio-33ac1c1e745694940450cb2fa05b596a9885d160943d6e41d82b80d7e428f21a WatchSource:0}: Error finding container 33ac1c1e745694940450cb2fa05b596a9885d160943d6e41d82b80d7e428f21a: Status 404 returned error can't find the container with id 33ac1c1e745694940450cb2fa05b596a9885d160943d6e41d82b80d7e428f21a Apr 22 13:22:09.548622 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.548597 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-smwrp"] Apr 22 13:22:09.553987 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:09.553959 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod512808ab_fbb1_46eb_86e9_a0fb60d532db.slice/crio-ebb21db6fe90a5db10595ac5cccf2fb8384d01d08525b1d1fdb37dc0c637c276 WatchSource:0}: Error finding container ebb21db6fe90a5db10595ac5cccf2fb8384d01d08525b1d1fdb37dc0c637c276: Status 404 returned error can't find the container with id ebb21db6fe90a5db10595ac5cccf2fb8384d01d08525b1d1fdb37dc0c637c276 Apr 22 13:22:09.758492 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.758164 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xg89z" event={"ID":"52d135c0-aa2e-4533-986a-5002ff68781d","Type":"ContainerStarted","Data":"33ac1c1e745694940450cb2fa05b596a9885d160943d6e41d82b80d7e428f21a"} Apr 22 13:22:09.759818 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.759783 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" event={"ID":"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f","Type":"ContainerStarted","Data":"2d9eea46fbf50a09be71217d5dfd929a8e1a662789241bb1407d4c7965d643c1"} Apr 22 13:22:09.759946 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.759829 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" event={"ID":"d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f","Type":"ContainerStarted","Data":"e7ffc9e95ff0b34b8001907ee5a73cf0d9613f945f25e09a8389bc9ea7d94ef8"} Apr 22 13:22:09.759946 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.759934 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:09.761071 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.761040 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-smwrp" event={"ID":"512808ab-fbb1-46eb-86e9-a0fb60d532db","Type":"ContainerStarted","Data":"5d8976704579b8d60aed040674415c2e448def0a617eb05d95e20a212b98dc2f"} Apr 22 13:22:09.761071 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.761071 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-smwrp" event={"ID":"512808ab-fbb1-46eb-86e9-a0fb60d532db","Type":"ContainerStarted","Data":"ebb21db6fe90a5db10595ac5cccf2fb8384d01d08525b1d1fdb37dc0c637c276"} Apr 22 13:22:09.762000 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.761979 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-q8tfx" event={"ID":"82906f03-65d1-460f-b410-9f8526350920","Type":"ContainerStarted","Data":"e0035c031ffcf0c9c07e5797a5688f20a342d87a7826fdeca639eea2b03b7e2d"} Apr 22 13:22:09.782070 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:09.782028 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" podStartSLOduration=6.7820151939999995 podStartE2EDuration="6.782015194s" podCreationTimestamp="2026-04-22 13:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 13:22:09.781903966 +0000 UTC m=+47.718290832" watchObservedRunningTime="2026-04-22 13:22:09.782015194 +0000 UTC m=+47.718402056" Apr 22 13:22:10.600372 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.600333 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:10.600554 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.600384 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:10.601068 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.600333 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:10.604138 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.604113 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 22 13:22:10.604406 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.604390 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-8cjvt\"" Apr 22 13:22:10.604616 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.604600 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-bxmrr\"" Apr 22 13:22:10.604791 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.604775 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 22 13:22:10.604922 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.604109 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 22 13:22:10.604995 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:10.604978 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 22 13:22:12.770191 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:12.770003 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xg89z" event={"ID":"52d135c0-aa2e-4533-986a-5002ff68781d","Type":"ContainerStarted","Data":"c4187a1e86b6d07472a98b98d623ea8148f1c2be22a2c91c99ef19f82f10212a"} Apr 22 13:22:12.771588 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:12.771568 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-smwrp" event={"ID":"512808ab-fbb1-46eb-86e9-a0fb60d532db","Type":"ContainerStarted","Data":"efe95930ff7da68210410f97586eb23f9d78a028888de494b03fc13c82aa9527"} Apr 22 13:22:12.772817 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:12.772788 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-q8tfx" event={"ID":"82906f03-65d1-460f-b410-9f8526350920","Type":"ContainerStarted","Data":"20efbcf2182483fc6c2df74ca45a71768c0254aae9889bf81485a3206e04483d"} Apr 22 13:22:12.772906 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:12.772820 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-q8tfx" event={"ID":"82906f03-65d1-460f-b410-9f8526350920","Type":"ContainerStarted","Data":"9effc69187a8cb62639d5344dbef38169ad27bd342bab4f5fe79119aaf3ca81a"} Apr 22 13:22:12.772952 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:12.772934 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:12.785645 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:12.785476 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-xg89z" podStartSLOduration=2.403104681 podStartE2EDuration="4.785464387s" podCreationTimestamp="2026-04-22 13:22:08 +0000 UTC" firstStartedPulling="2026-04-22 13:22:09.529956474 +0000 UTC m=+47.466343319" lastFinishedPulling="2026-04-22 13:22:11.912316184 +0000 UTC m=+49.848703025" observedRunningTime="2026-04-22 13:22:12.785301228 +0000 UTC m=+50.721688093" watchObservedRunningTime="2026-04-22 13:22:12.785464387 +0000 UTC m=+50.721851251" Apr 22 13:22:12.803874 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:12.803832 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-q8tfx" podStartSLOduration=2.410934591 podStartE2EDuration="4.803819796s" podCreationTimestamp="2026-04-22 13:22:08 +0000 UTC" firstStartedPulling="2026-04-22 13:22:09.514680354 +0000 UTC m=+47.451067209" lastFinishedPulling="2026-04-22 13:22:11.907565565 +0000 UTC m=+49.843952414" observedRunningTime="2026-04-22 13:22:12.802354899 +0000 UTC m=+50.738741765" watchObservedRunningTime="2026-04-22 13:22:12.803819796 +0000 UTC m=+50.740206650" Apr 22 13:22:14.780065 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:14.780026 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-smwrp" event={"ID":"512808ab-fbb1-46eb-86e9-a0fb60d532db","Type":"ContainerStarted","Data":"30612e92997ddfb92aa134da3e371d12c9e3f09090555ac4ec1fee34100eb11e"} Apr 22 13:22:16.186910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.186861 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-smwrp" podStartSLOduration=4.032503895 podStartE2EDuration="8.186843079s" podCreationTimestamp="2026-04-22 13:22:08 +0000 UTC" firstStartedPulling="2026-04-22 13:22:09.644173525 +0000 UTC m=+47.580560369" lastFinishedPulling="2026-04-22 13:22:13.798512702 +0000 UTC m=+51.734899553" observedRunningTime="2026-04-22 13:22:14.7988394 +0000 UTC m=+52.735226265" watchObservedRunningTime="2026-04-22 13:22:16.186843079 +0000 UTC m=+54.123229941" Apr 22 13:22:16.187408 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.187392 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-rz469"] Apr 22 13:22:16.222053 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.222024 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-rz469"] Apr 22 13:22:16.222053 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.222049 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-d5jgb"] Apr 22 13:22:16.222211 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.222165 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.224518 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.224499 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 22 13:22:16.225471 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.225456 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-dockercfg-27c5j\"" Apr 22 13:22:16.225544 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.225477 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 22 13:22:16.225544 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.225493 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 22 13:22:16.225544 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.225505 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-kube-rbac-proxy-config\"" Apr 22 13:22:16.226598 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.226581 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"openshift-state-metrics-tls\"" Apr 22 13:22:16.240557 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.240538 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.243129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.242942 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-x2x6b\"" Apr 22 13:22:16.243129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.242989 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 22 13:22:16.243282 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.243218 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 22 13:22:16.244633 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.243550 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 22 13:22:16.303567 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303547 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.303685 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303576 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-accelerators-collector-config\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303685 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303594 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/33daaf5e-222e-4197-a370-52af9fb7f55c-metrics-client-ca\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303685 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303633 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-tls\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303685 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303657 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgqwc\" (UniqueName: \"kubernetes.io/projected/33daaf5e-222e-4197-a370-52af9fb7f55c-kube-api-access-zgqwc\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303685 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303676 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-root\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303855 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303693 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kc9j\" (UniqueName: \"kubernetes.io/projected/9d530630-78b4-49c7-9015-fc89b9f15391-kube-api-access-5kc9j\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.303855 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303773 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-textfile\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303855 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303828 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9d530630-78b4-49c7-9015-fc89b9f15391-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.303855 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303848 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303988 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303896 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.303988 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303921 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-sys\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.303988 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.303936 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-wtmp\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.405239 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405209 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/33daaf5e-222e-4197-a370-52af9fb7f55c-metrics-client-ca\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.405384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405258 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-tls\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.405384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405275 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zgqwc\" (UniqueName: \"kubernetes.io/projected/33daaf5e-222e-4197-a370-52af9fb7f55c-kube-api-access-zgqwc\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.405384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405292 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-root\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.405384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405340 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-root\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.405605 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:16.405394 2579 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Apr 22 13:22:16.405605 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:16.405465 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-tls podName:33daaf5e-222e-4197-a370-52af9fb7f55c nodeName:}" failed. No retries permitted until 2026-04-22 13:22:16.905449006 +0000 UTC m=+54.841835846 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-tls") pod "node-exporter-d5jgb" (UID: "33daaf5e-222e-4197-a370-52af9fb7f55c") : secret "node-exporter-tls" not found Apr 22 13:22:16.405605 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405461 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5kc9j\" (UniqueName: \"kubernetes.io/projected/9d530630-78b4-49c7-9015-fc89b9f15391-kube-api-access-5kc9j\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.405605 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405525 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-textfile\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.405605 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405565 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9d530630-78b4-49c7-9015-fc89b9f15391-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.405605 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405588 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405625 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405666 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-sys\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405690 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-wtmp\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405716 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405743 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-accelerators-collector-config\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.405833 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-wtmp\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:16.405919 2579 secret.go:189] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Apr 22 13:22:16.406045 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:16.405965 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-tls podName:9d530630-78b4-49c7-9015-fc89b9f15391 nodeName:}" failed. No retries permitted until 2026-04-22 13:22:16.905950372 +0000 UTC m=+54.842337213 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-tls") pod "openshift-state-metrics-9d44df66c-rz469" (UID: "9d530630-78b4-49c7-9015-fc89b9f15391") : secret "openshift-state-metrics-tls" not found Apr 22 13:22:16.406421 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.406320 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9d530630-78b4-49c7-9015-fc89b9f15391-metrics-client-ca\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.406421 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.406369 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33daaf5e-222e-4197-a370-52af9fb7f55c-sys\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.409506 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.409486 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.412424 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.412403 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-textfile\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.412523 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.412505 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/33daaf5e-222e-4197-a370-52af9fb7f55c-metrics-client-ca\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.412630 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.412609 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.412672 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.412636 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-accelerators-collector-config\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.418779 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.418756 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgqwc\" (UniqueName: \"kubernetes.io/projected/33daaf5e-222e-4197-a370-52af9fb7f55c-kube-api-access-zgqwc\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.420536 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.420514 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kc9j\" (UniqueName: \"kubernetes.io/projected/9d530630-78b4-49c7-9015-fc89b9f15391-kube-api-access-5kc9j\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.908325 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.908293 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:16.908486 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.908340 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-tls\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.910409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.910385 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/33daaf5e-222e-4197-a370-52af9fb7f55c-node-exporter-tls\") pod \"node-exporter-d5jgb\" (UID: \"33daaf5e-222e-4197-a370-52af9fb7f55c\") " pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:16.910656 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:16.910636 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d530630-78b4-49c7-9015-fc89b9f15391-openshift-state-metrics-tls\") pod \"openshift-state-metrics-9d44df66c-rz469\" (UID: \"9d530630-78b4-49c7-9015-fc89b9f15391\") " pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:17.131312 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.131276 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" Apr 22 13:22:17.151843 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.151810 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-d5jgb" Apr 22 13:22:17.161742 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:17.161715 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33daaf5e_222e_4197_a370_52af9fb7f55c.slice/crio-a8a79725bc715113ebb701422e170d97f062abe7056502c796ea0720604461ec WatchSource:0}: Error finding container a8a79725bc715113ebb701422e170d97f062abe7056502c796ea0720604461ec: Status 404 returned error can't find the container with id a8a79725bc715113ebb701422e170d97f062abe7056502c796ea0720604461ec Apr 22 13:22:17.247140 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.247112 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-9d44df66c-rz469"] Apr 22 13:22:17.250041 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:17.250012 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d530630_78b4_49c7_9015_fc89b9f15391.slice/crio-12e35fa34c9288631c9d0af3097417e2b1b935dd5235ab8b71856ae33f1c5389 WatchSource:0}: Error finding container 12e35fa34c9288631c9d0af3097417e2b1b935dd5235ab8b71856ae33f1c5389: Status 404 returned error can't find the container with id 12e35fa34c9288631c9d0af3097417e2b1b935dd5235ab8b71856ae33f1c5389 Apr 22 13:22:17.306072 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.306044 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:22:17.336576 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.336551 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:22:17.336712 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.336699 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.339293 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339270 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Apr 22 13:22:17.339293 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339284 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-r4gw2\"" Apr 22 13:22:17.339448 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339315 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Apr 22 13:22:17.339448 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339340 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Apr 22 13:22:17.339448 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339423 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Apr 22 13:22:17.339608 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339591 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Apr 22 13:22:17.339861 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339843 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Apr 22 13:22:17.339929 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339870 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Apr 22 13:22:17.339929 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.339891 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Apr 22 13:22:17.340690 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.340675 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Apr 22 13:22:17.412245 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412223 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412310 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412254 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-config-out\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412310 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412274 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412310 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412293 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r67gw\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-kube-api-access-r67gw\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412481 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412401 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412481 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412461 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412492 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412528 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-config-volume\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412569 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412712 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412626 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412712 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412651 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412712 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412683 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-tls-assets\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.412804 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.412731 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-web-config\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.513738 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.513710 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.513843 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.513760 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.513843 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.513805 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.513843 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.513833 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-config-volume\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514036 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514018 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514130 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514064 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514130 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514102 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514133 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514136 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-tls-assets\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514230 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-web-config\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:17.514249 2579 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle podName:a758203c-d315-4a59-b93f-4fa9e34884b6 nodeName:}" failed. No retries permitted until 2026-04-22 13:22:18.014230654 +0000 UTC m=+55.950617509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "alertmanager-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle") pod "alertmanager-main-0" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6") : configmap references non-existent config key: ca-bundle.crt Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514300 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514332 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-config-out\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514362 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.514834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.514387 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r67gw\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-kube-api-access-r67gw\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.515518 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.515414 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.517616 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.517590 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.517702 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.517680 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.517930 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.517889 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-config-out\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.518330 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.518286 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.518465 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.518423 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-config-volume\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.518708 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.518646 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.518851 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.518830 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-tls-assets\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.518936 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.518922 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.519527 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.519510 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-web-config\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.522835 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.522819 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r67gw\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-kube-api-access-r67gw\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:17.790644 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.790606 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-d5jgb" event={"ID":"33daaf5e-222e-4197-a370-52af9fb7f55c","Type":"ContainerStarted","Data":"a8a79725bc715113ebb701422e170d97f062abe7056502c796ea0720604461ec"} Apr 22 13:22:17.792505 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.792470 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" event={"ID":"9d530630-78b4-49c7-9015-fc89b9f15391","Type":"ContainerStarted","Data":"88f737f9a171f6114f611e3985520a8f5338fdfab70751524012eda9250cde9a"} Apr 22 13:22:17.792505 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.792500 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" event={"ID":"9d530630-78b4-49c7-9015-fc89b9f15391","Type":"ContainerStarted","Data":"df4159bc84015b8a397b04d6f549847851ecbbf4dd4d2333d4dda85ba08aae6d"} Apr 22 13:22:17.792652 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:17.792513 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" event={"ID":"9d530630-78b4-49c7-9015-fc89b9f15391","Type":"ContainerStarted","Data":"12e35fa34c9288631c9d0af3097417e2b1b935dd5235ab8b71856ae33f1c5389"} Apr 22 13:22:18.019299 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.019275 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:18.020039 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.020018 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:18.245178 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.245154 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:22:18.374859 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.374780 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:22:18.376919 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:18.376886 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda758203c_d315_4a59_b93f_4fa9e34884b6.slice/crio-7f5fa518133c2c9bf9cee16f9c05a0c6f4aa65f714a3955196f0feb6564fefe2 WatchSource:0}: Error finding container 7f5fa518133c2c9bf9cee16f9c05a0c6f4aa65f714a3955196f0feb6564fefe2: Status 404 returned error can't find the container with id 7f5fa518133c2c9bf9cee16f9c05a0c6f4aa65f714a3955196f0feb6564fefe2 Apr 22 13:22:18.796197 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.796162 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerStarted","Data":"7f5fa518133c2c9bf9cee16f9c05a0c6f4aa65f714a3955196f0feb6564fefe2"} Apr 22 13:22:18.797548 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.797523 2579 generic.go:358] "Generic (PLEG): container finished" podID="33daaf5e-222e-4197-a370-52af9fb7f55c" containerID="4ea74152c3c8cf9291fb20bebbbf65be96c0dcbd7687fea943844860c8c51144" exitCode=0 Apr 22 13:22:18.797669 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.797600 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-d5jgb" event={"ID":"33daaf5e-222e-4197-a370-52af9fb7f55c","Type":"ContainerDied","Data":"4ea74152c3c8cf9291fb20bebbbf65be96c0dcbd7687fea943844860c8c51144"} Apr 22 13:22:18.799330 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.799310 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" event={"ID":"9d530630-78b4-49c7-9015-fc89b9f15391","Type":"ContainerStarted","Data":"f8b6b2ac35864ad870d929d9ef32a39be33d0bc4b8a208cb4a9934ebdbaddf06"} Apr 22 13:22:18.847778 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:18.847732 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-9d44df66c-rz469" podStartSLOduration=1.67767471 podStartE2EDuration="2.847715728s" podCreationTimestamp="2026-04-22 13:22:16 +0000 UTC" firstStartedPulling="2026-04-22 13:22:17.503539494 +0000 UTC m=+55.439926338" lastFinishedPulling="2026-04-22 13:22:18.673580511 +0000 UTC m=+56.609967356" observedRunningTime="2026-04-22 13:22:18.84637791 +0000 UTC m=+56.782764774" watchObservedRunningTime="2026-04-22 13:22:18.847715728 +0000 UTC m=+56.784102594" Apr 22 13:22:19.803261 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:19.803228 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-d5jgb" event={"ID":"33daaf5e-222e-4197-a370-52af9fb7f55c","Type":"ContainerStarted","Data":"22d7a6e869ee86a32b24fc715a34e2deae454195d496812860798978c511c167"} Apr 22 13:22:19.803657 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:19.803270 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-d5jgb" event={"ID":"33daaf5e-222e-4197-a370-52af9fb7f55c","Type":"ContainerStarted","Data":"763b52fbe23200c177cdaedfa7ed41db574fe098a507578e860810c59655642d"} Apr 22 13:22:19.804487 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:19.804464 2579 generic.go:358] "Generic (PLEG): container finished" podID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerID="b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545" exitCode=0 Apr 22 13:22:19.804580 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:19.804544 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545"} Apr 22 13:22:19.823579 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:19.823541 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-d5jgb" podStartSLOduration=2.969314604 podStartE2EDuration="3.823528564s" podCreationTimestamp="2026-04-22 13:22:16 +0000 UTC" firstStartedPulling="2026-04-22 13:22:17.163231924 +0000 UTC m=+55.099618769" lastFinishedPulling="2026-04-22 13:22:18.017445873 +0000 UTC m=+55.953832729" observedRunningTime="2026-04-22 13:22:19.822216046 +0000 UTC m=+57.758602943" watchObservedRunningTime="2026-04-22 13:22:19.823528564 +0000 UTC m=+57.759915473" Apr 22 13:22:20.350267 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.350237 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-54b485d8d6-8jlp9"] Apr 22 13:22:20.353572 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.353555 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.356275 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.356245 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 22 13:22:20.356275 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.356253 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 22 13:22:20.357736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.357536 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-rgk2k\"" Apr 22 13:22:20.357736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.357546 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 22 13:22:20.357736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.357563 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 22 13:22:20.357736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.357607 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 22 13:22:20.357736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.357541 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 22 13:22:20.358022 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.357886 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 22 13:22:20.362165 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.362149 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 22 13:22:20.362395 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.362380 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54b485d8d6-8jlp9"] Apr 22 13:22:20.440152 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.440127 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-oauth-config\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.440253 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.440157 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-trusted-ca-bundle\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.440253 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.440174 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-oauth-serving-cert\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.440253 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.440233 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-serving-cert\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.440253 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.440248 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r4xb\" (UniqueName: \"kubernetes.io/projected/982000d1-4a44-402b-bd86-d5df1f823d5c-kube-api-access-5r4xb\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.440390 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.440269 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-console-config\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.440390 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.440330 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-service-ca\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.532381 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.532356 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-757fc4c546-cgzdq"] Apr 22 13:22:20.535343 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.535327 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.537632 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.537611 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-dockercfg-7nsq4\"" Apr 22 13:22:20.537746 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.537724 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-client-certs\"" Apr 22 13:22:20.537818 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.537788 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-server-audit-profiles\"" Apr 22 13:22:20.537966 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.537951 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-4c7uci4sbmpii\"" Apr 22 13:22:20.538077 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.538058 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-tls\"" Apr 22 13:22:20.538150 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.538135 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kubelet-serving-ca-bundle\"" Apr 22 13:22:20.540631 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.540613 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-serving-cert\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.540736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.540641 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5r4xb\" (UniqueName: \"kubernetes.io/projected/982000d1-4a44-402b-bd86-d5df1f823d5c-kube-api-access-5r4xb\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.540736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.540667 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-console-config\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.540736 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.540702 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-service-ca\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.540907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.540782 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-oauth-config\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.540907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.540813 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-trusted-ca-bundle\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.540907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.540840 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-oauth-serving-cert\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.542508 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.541785 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-console-config\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.542508 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.541914 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-trusted-ca-bundle\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.542508 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.542141 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-service-ca\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.542508 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.542205 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-oauth-serving-cert\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.543604 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.543556 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-serving-cert\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.543828 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.543804 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-oauth-config\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.544900 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.544881 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-757fc4c546-cgzdq"] Apr 22 13:22:20.548145 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.548119 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r4xb\" (UniqueName: \"kubernetes.io/projected/982000d1-4a44-402b-bd86-d5df1f823d5c-kube-api-access-5r4xb\") pod \"console-54b485d8d6-8jlp9\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.642189 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.642124 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl67m\" (UniqueName: \"kubernetes.io/projected/e6e3682a-be46-4080-8f7f-01f99989110c-kube-api-access-bl67m\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.642189 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.642172 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-secret-metrics-server-tls\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.642356 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.642198 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-client-ca-bundle\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.642356 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.642241 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-secret-metrics-server-client-certs\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.642356 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.642310 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e6e3682a-be46-4080-8f7f-01f99989110c-metrics-server-audit-profiles\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.642356 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.642345 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6e3682a-be46-4080-8f7f-01f99989110c-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.642520 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.642367 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e6e3682a-be46-4080-8f7f-01f99989110c-audit-log\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.663122 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.663105 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:20.723047 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.723022 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqksf" Apr 22 13:22:20.742893 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.742867 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-secret-metrics-server-tls\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.743003 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.742900 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-client-ca-bundle\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.743724 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.743139 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-secret-metrics-server-client-certs\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.743724 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.743233 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e6e3682a-be46-4080-8f7f-01f99989110c-metrics-server-audit-profiles\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.743724 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.743274 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6e3682a-be46-4080-8f7f-01f99989110c-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.743724 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.743306 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e6e3682a-be46-4080-8f7f-01f99989110c-audit-log\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.743724 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.743390 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bl67m\" (UniqueName: \"kubernetes.io/projected/e6e3682a-be46-4080-8f7f-01f99989110c-kube-api-access-bl67m\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.745326 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.744491 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e6e3682a-be46-4080-8f7f-01f99989110c-metrics-server-audit-profiles\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.745326 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.744809 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e6e3682a-be46-4080-8f7f-01f99989110c-audit-log\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.745326 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.745281 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6e3682a-be46-4080-8f7f-01f99989110c-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.745858 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.745834 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-secret-metrics-server-client-certs\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.745945 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.745915 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-secret-metrics-server-tls\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.746637 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.746606 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e3682a-be46-4080-8f7f-01f99989110c-client-ca-bundle\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.754858 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.754835 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl67m\" (UniqueName: \"kubernetes.io/projected/e6e3682a-be46-4080-8f7f-01f99989110c-kube-api-access-bl67m\") pod \"metrics-server-757fc4c546-cgzdq\" (UID: \"e6e3682a-be46-4080-8f7f-01f99989110c\") " pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.781670 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.781649 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54b485d8d6-8jlp9"] Apr 22 13:22:20.784458 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:20.784408 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod982000d1_4a44_402b_bd86_d5df1f823d5c.slice/crio-a1117057e5e5dc1f377cc884323e7ac2c41e04ed5952b7f01fd05912f62bdb72 WatchSource:0}: Error finding container a1117057e5e5dc1f377cc884323e7ac2c41e04ed5952b7f01fd05912f62bdb72: Status 404 returned error can't find the container with id a1117057e5e5dc1f377cc884323e7ac2c41e04ed5952b7f01fd05912f62bdb72 Apr 22 13:22:20.808956 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.808924 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54b485d8d6-8jlp9" event={"ID":"982000d1-4a44-402b-bd86-d5df1f823d5c","Type":"ContainerStarted","Data":"a1117057e5e5dc1f377cc884323e7ac2c41e04ed5952b7f01fd05912f62bdb72"} Apr 22 13:22:20.844951 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.844927 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:20.964350 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:20.964324 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-757fc4c546-cgzdq"] Apr 22 13:22:20.967264 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:20.967239 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6e3682a_be46_4080_8f7f_01f99989110c.slice/crio-61ac1f1e1929548830062ad1e1c2f1dd9b842a41cc4ff3b9a8766d68cdbf98c3 WatchSource:0}: Error finding container 61ac1f1e1929548830062ad1e1c2f1dd9b842a41cc4ff3b9a8766d68cdbf98c3: Status 404 returned error can't find the container with id 61ac1f1e1929548830062ad1e1c2f1dd9b842a41cc4ff3b9a8766d68cdbf98c3 Apr 22 13:22:21.815086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:21.815054 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerStarted","Data":"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277"} Apr 22 13:22:21.816358 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:21.816332 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" event={"ID":"e6e3682a-be46-4080-8f7f-01f99989110c","Type":"ContainerStarted","Data":"61ac1f1e1929548830062ad1e1c2f1dd9b842a41cc4ff3b9a8766d68cdbf98c3"} Apr 22 13:22:22.435645 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.435514 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:22:22.439364 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.439322 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.442996 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.442660 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Apr 22 13:22:22.442996 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.442824 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Apr 22 13:22:22.442996 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.442887 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-2dq2oeagafuup\"" Apr 22 13:22:22.442996 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.442994 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Apr 22 13:22:22.443391 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.443031 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Apr 22 13:22:22.443391 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.443341 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-llln8\"" Apr 22 13:22:22.443850 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.443671 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Apr 22 13:22:22.443850 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.443800 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Apr 22 13:22:22.444086 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.443947 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Apr 22 13:22:22.444499 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.444205 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Apr 22 13:22:22.444499 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.444492 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Apr 22 13:22:22.445540 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.444763 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Apr 22 13:22:22.445540 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.444993 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Apr 22 13:22:22.448670 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.447778 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Apr 22 13:22:22.454562 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.454207 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:22:22.564089 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564057 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564232 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564121 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564232 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564151 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564232 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564183 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config-out\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564232 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564213 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564260 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564285 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564316 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9565\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-kube-api-access-t9565\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564345 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564376 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-web-config\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564640 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564413 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564640 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564456 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564640 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564486 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564640 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564510 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564640 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564566 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564640 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564608 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564901 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564661 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.564901 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.564722 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.665755 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.665724 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.665921 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.665761 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.665921 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.665890 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.665921 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.665917 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666085 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.665955 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config-out\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666085 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666005 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666085 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666042 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666085 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666069 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666273 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666097 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-t9565\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-kube-api-access-t9565\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666273 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666122 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666273 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666154 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-web-config\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666273 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666191 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666496 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666321 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666496 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666355 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666496 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666379 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666496 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666407 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666496 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666454 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666735 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666505 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666735 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666542 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.666735 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.666644 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.667047 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.667025 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.668729 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.668707 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config-out\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.669188 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.668949 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Apr 22 13:22:22.669188 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.668974 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Apr 22 13:22:22.669188 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.668949 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Apr 22 13:22:22.669188 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.668949 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Apr 22 13:22:22.669188 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.669009 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Apr 22 13:22:22.669188 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.669031 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Apr 22 13:22:22.669540 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.669516 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Apr 22 13:22:22.669746 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.669729 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-2dq2oeagafuup\"" Apr 22 13:22:22.669829 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.669796 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Apr 22 13:22:22.669921 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.669897 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Apr 22 13:22:22.670029 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.669960 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Apr 22 13:22:22.671942 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.671924 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Apr 22 13:22:22.675558 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.675540 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Apr 22 13:22:22.676292 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.676276 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9565\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-kube-api-access-t9565\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.679456 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.677860 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.679456 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.679307 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.679456 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.679343 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.679657 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.679490 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-web-config\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.679869 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.679807 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.680170 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.680147 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.680523 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.680501 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.680824 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.680797 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.682061 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.682036 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.682472 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.682418 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.682677 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.682658 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.682738 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.682711 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.682861 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.682839 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.756543 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.756515 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-llln8\"" Apr 22 13:22:22.764522 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.764499 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:22.778474 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.778451 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-q8tfx" Apr 22 13:22:22.822698 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.822665 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerStarted","Data":"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56"} Apr 22 13:22:22.823106 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.822706 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerStarted","Data":"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3"} Apr 22 13:22:22.823106 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.822724 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerStarted","Data":"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a"} Apr 22 13:22:22.823106 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:22.822738 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerStarted","Data":"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f"} Apr 22 13:22:24.155766 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.155674 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:22:24.158778 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:24.158744 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d9eec3c_17ba_4fcd_ac61_5fdbc1dc7ed7.slice/crio-2da4f9c3ab245741675f2b1ebb0e5aeddb654c8bbd4f6a041a6348dc191ae6b1 WatchSource:0}: Error finding container 2da4f9c3ab245741675f2b1ebb0e5aeddb654c8bbd4f6a041a6348dc191ae6b1: Status 404 returned error can't find the container with id 2da4f9c3ab245741675f2b1ebb0e5aeddb654c8bbd4f6a041a6348dc191ae6b1 Apr 22 13:22:24.834905 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.834870 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerStarted","Data":"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db"} Apr 22 13:22:24.836296 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.836265 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" event={"ID":"e6e3682a-be46-4080-8f7f-01f99989110c","Type":"ContainerStarted","Data":"bf085ace6ee47e0c9dabad5a7fd2c012c207530f01f16a85ce4326d4f927545f"} Apr 22 13:22:24.837500 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.837477 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54b485d8d6-8jlp9" event={"ID":"982000d1-4a44-402b-bd86-d5df1f823d5c","Type":"ContainerStarted","Data":"72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc"} Apr 22 13:22:24.838680 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.838655 2579 generic.go:358] "Generic (PLEG): container finished" podID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerID="6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8" exitCode=0 Apr 22 13:22:24.838758 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.838700 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8"} Apr 22 13:22:24.838758 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.838720 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerStarted","Data":"2da4f9c3ab245741675f2b1ebb0e5aeddb654c8bbd4f6a041a6348dc191ae6b1"} Apr 22 13:22:24.861899 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.861859 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.183648421 podStartE2EDuration="7.861844563s" podCreationTimestamp="2026-04-22 13:22:17 +0000 UTC" firstStartedPulling="2026-04-22 13:22:18.379148759 +0000 UTC m=+56.315535600" lastFinishedPulling="2026-04-22 13:22:24.057344895 +0000 UTC m=+61.993731742" observedRunningTime="2026-04-22 13:22:24.860615828 +0000 UTC m=+62.797002691" watchObservedRunningTime="2026-04-22 13:22:24.861844563 +0000 UTC m=+62.798231423" Apr 22 13:22:24.882068 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.882028 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54b485d8d6-8jlp9" podStartSLOduration=1.664206059 podStartE2EDuration="4.882013891s" podCreationTimestamp="2026-04-22 13:22:20 +0000 UTC" firstStartedPulling="2026-04-22 13:22:20.786222176 +0000 UTC m=+58.722609020" lastFinishedPulling="2026-04-22 13:22:24.004030012 +0000 UTC m=+61.940416852" observedRunningTime="2026-04-22 13:22:24.88118009 +0000 UTC m=+62.817566954" watchObservedRunningTime="2026-04-22 13:22:24.882013891 +0000 UTC m=+62.818400754" Apr 22 13:22:24.900682 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:24.900646 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" podStartSLOduration=1.867713213 podStartE2EDuration="4.900633517s" podCreationTimestamp="2026-04-22 13:22:20 +0000 UTC" firstStartedPulling="2026-04-22 13:22:20.969808032 +0000 UTC m=+58.906194891" lastFinishedPulling="2026-04-22 13:22:24.002728354 +0000 UTC m=+61.939115195" observedRunningTime="2026-04-22 13:22:24.899560664 +0000 UTC m=+62.835947538" watchObservedRunningTime="2026-04-22 13:22:24.900633517 +0000 UTC m=+62.837020379" Apr 22 13:22:27.561098 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:27.561055 2579 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-54b485d8d6-8jlp9"] Apr 22 13:22:28.323079 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.323036 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:28.323260 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.323115 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:28.325940 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.325919 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 22 13:22:28.325989 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.325964 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 22 13:22:28.336318 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.336294 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dec8b65-6274-44b1-a169-ad20045acd1a-metrics-certs\") pod \"network-metrics-daemon-8vcg7\" (UID: \"5dec8b65-6274-44b1-a169-ad20045acd1a\") " pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:28.336703 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.336687 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 22 13:22:28.346989 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.346963 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbhdt\" (UniqueName: \"kubernetes.io/projected/4829e038-c6a2-4934-8182-fae81fb059a8-kube-api-access-zbhdt\") pod \"network-check-target-tmp8r\" (UID: \"4829e038-c6a2-4934-8182-fae81fb059a8\") " pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:28.423601 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.423564 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:28.426074 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.426055 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 22 13:22:28.436523 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.436495 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/6560c958-0928-4fa1-8691-8fe812c8d951-original-pull-secret\") pod \"global-pull-secret-syncer-c7j5x\" (UID: \"6560c958-0928-4fa1-8691-8fe812c8d951\") " pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:28.618994 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.618908 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-8cjvt\"" Apr 22 13:22:28.625326 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.625298 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-bxmrr\"" Apr 22 13:22:28.627297 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.627276 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:28.629011 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.628989 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-c7j5x" Apr 22 13:22:28.633665 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.633645 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8vcg7" Apr 22 13:22:28.823485 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.823398 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-tmp8r"] Apr 22 13:22:28.854082 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.854045 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerStarted","Data":"ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb"} Apr 22 13:22:28.854082 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.854083 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerStarted","Data":"2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8"} Apr 22 13:22:28.855223 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:28.855198 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-tmp8r" event={"ID":"4829e038-c6a2-4934-8182-fae81fb059a8","Type":"ContainerStarted","Data":"f9519953d1379bdb2854fdc4868d907dd25b930acf1bd771d282dcea9e0fa59e"} Apr 22 13:22:29.057177 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:29.056803 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8vcg7"] Apr 22 13:22:29.059149 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:29.059118 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-c7j5x"] Apr 22 13:22:29.060497 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:29.060467 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5dec8b65_6274_44b1_a169_ad20045acd1a.slice/crio-e2b839c2b109fcc1d81623120a09d889b40036d93a26a1e454e7766202489d62 WatchSource:0}: Error finding container e2b839c2b109fcc1d81623120a09d889b40036d93a26a1e454e7766202489d62: Status 404 returned error can't find the container with id e2b839c2b109fcc1d81623120a09d889b40036d93a26a1e454e7766202489d62 Apr 22 13:22:29.062269 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:22:29.062241 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6560c958_0928_4fa1_8691_8fe812c8d951.slice/crio-022373abd96ee2bcdb35e174418e1dd3354530b13613c7983acc65a42637d9f8 WatchSource:0}: Error finding container 022373abd96ee2bcdb35e174418e1dd3354530b13613c7983acc65a42637d9f8: Status 404 returned error can't find the container with id 022373abd96ee2bcdb35e174418e1dd3354530b13613c7983acc65a42637d9f8 Apr 22 13:22:29.860185 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:29.860142 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-c7j5x" event={"ID":"6560c958-0928-4fa1-8691-8fe812c8d951","Type":"ContainerStarted","Data":"022373abd96ee2bcdb35e174418e1dd3354530b13613c7983acc65a42637d9f8"} Apr 22 13:22:29.861398 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:29.861303 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8vcg7" event={"ID":"5dec8b65-6274-44b1-a169-ad20045acd1a","Type":"ContainerStarted","Data":"e2b839c2b109fcc1d81623120a09d889b40036d93a26a1e454e7766202489d62"} Apr 22 13:22:30.664374 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:30.664332 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:30.771037 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:30.770778 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-756c9666d4-q6jfg" Apr 22 13:22:30.869937 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:30.869848 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerStarted","Data":"2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66"} Apr 22 13:22:30.869937 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:30.869900 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerStarted","Data":"0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554"} Apr 22 13:22:30.869937 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:30.869914 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerStarted","Data":"2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e"} Apr 22 13:22:33.883812 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:33.883723 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerStarted","Data":"3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751"} Apr 22 13:22:33.921165 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:33.920683 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=6.381958432 podStartE2EDuration="11.92064843s" podCreationTimestamp="2026-04-22 13:22:22 +0000 UTC" firstStartedPulling="2026-04-22 13:22:24.839728876 +0000 UTC m=+62.776115730" lastFinishedPulling="2026-04-22 13:22:30.37841887 +0000 UTC m=+68.314805728" observedRunningTime="2026-04-22 13:22:33.918689628 +0000 UTC m=+71.855076494" watchObservedRunningTime="2026-04-22 13:22:33.92064843 +0000 UTC m=+71.857035290" Apr 22 13:22:34.889223 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.889185 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8vcg7" event={"ID":"5dec8b65-6274-44b1-a169-ad20045acd1a","Type":"ContainerStarted","Data":"ebcfd972f3b062f7bf9783f3f42af9988d79c8de6b5d8aefba11fa01305bbbfd"} Apr 22 13:22:34.889223 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.889227 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8vcg7" event={"ID":"5dec8b65-6274-44b1-a169-ad20045acd1a","Type":"ContainerStarted","Data":"86173b320507ff47cfd9c570a7dbd1133a6ec4666f13eee27727a97b468c46bf"} Apr 22 13:22:34.890500 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.890469 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-tmp8r" event={"ID":"4829e038-c6a2-4934-8182-fae81fb059a8","Type":"ContainerStarted","Data":"2b8f9babe7c751a299604f62e9fe52d162c978237d19fafe8250aa916d39f02f"} Apr 22 13:22:34.890646 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.890619 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:22:34.892040 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.892014 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-c7j5x" event={"ID":"6560c958-0928-4fa1-8691-8fe812c8d951","Type":"ContainerStarted","Data":"d0cc7f0d9dfb8387641bd02d2b484de8f6e691edf9ef65f33fcec72dfb480168"} Apr 22 13:22:34.906984 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.906941 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-8vcg7" podStartSLOduration=69.536470843 podStartE2EDuration="1m12.906926992s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:22:29.062734765 +0000 UTC m=+66.999121609" lastFinishedPulling="2026-04-22 13:22:32.43319091 +0000 UTC m=+70.369577758" observedRunningTime="2026-04-22 13:22:34.905540137 +0000 UTC m=+72.841927015" watchObservedRunningTime="2026-04-22 13:22:34.906926992 +0000 UTC m=+72.843313895" Apr 22 13:22:34.921037 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.920988 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-c7j5x" podStartSLOduration=66.124005959 podStartE2EDuration="1m10.920975936s" podCreationTimestamp="2026-04-22 13:21:24 +0000 UTC" firstStartedPulling="2026-04-22 13:22:29.064161595 +0000 UTC m=+67.000548439" lastFinishedPulling="2026-04-22 13:22:33.86113156 +0000 UTC m=+71.797518416" observedRunningTime="2026-04-22 13:22:34.920074004 +0000 UTC m=+72.856460866" watchObservedRunningTime="2026-04-22 13:22:34.920975936 +0000 UTC m=+72.857362840" Apr 22 13:22:34.935180 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:34.935128 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-tmp8r" podStartSLOduration=67.973376389 podStartE2EDuration="1m12.935111129s" podCreationTimestamp="2026-04-22 13:21:22 +0000 UTC" firstStartedPulling="2026-04-22 13:22:28.830377072 +0000 UTC m=+66.766763938" lastFinishedPulling="2026-04-22 13:22:33.792111821 +0000 UTC m=+71.728498678" observedRunningTime="2026-04-22 13:22:34.93422238 +0000 UTC m=+72.870609254" watchObservedRunningTime="2026-04-22 13:22:34.935111129 +0000 UTC m=+72.871497991" Apr 22 13:22:37.765214 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:37.765072 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:22:40.845895 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:40.845842 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:40.845895 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:40.845906 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:22:52.581314 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.581245 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-54b485d8d6-8jlp9" podUID="982000d1-4a44-402b-bd86-d5df1f823d5c" containerName="console" containerID="cri-o://72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc" gracePeriod=15 Apr 22 13:22:52.826815 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.826785 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-54b485d8d6-8jlp9_982000d1-4a44-402b-bd86-d5df1f823d5c/console/0.log" Apr 22 13:22:52.826941 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.826863 2579 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:52.836657 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.836600 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-console-config\") pod \"982000d1-4a44-402b-bd86-d5df1f823d5c\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " Apr 22 13:22:52.836657 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.836643 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-service-ca\") pod \"982000d1-4a44-402b-bd86-d5df1f823d5c\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " Apr 22 13:22:52.836835 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.836671 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-oauth-config\") pod \"982000d1-4a44-402b-bd86-d5df1f823d5c\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " Apr 22 13:22:52.836835 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.836693 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-oauth-serving-cert\") pod \"982000d1-4a44-402b-bd86-d5df1f823d5c\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " Apr 22 13:22:52.836835 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.836719 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-serving-cert\") pod \"982000d1-4a44-402b-bd86-d5df1f823d5c\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " Apr 22 13:22:52.836835 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.836738 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-trusted-ca-bundle\") pod \"982000d1-4a44-402b-bd86-d5df1f823d5c\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " Apr 22 13:22:52.836835 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.836760 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r4xb\" (UniqueName: \"kubernetes.io/projected/982000d1-4a44-402b-bd86-d5df1f823d5c-kube-api-access-5r4xb\") pod \"982000d1-4a44-402b-bd86-d5df1f823d5c\" (UID: \"982000d1-4a44-402b-bd86-d5df1f823d5c\") " Apr 22 13:22:52.837146 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.837112 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-console-config" (OuterVolumeSpecName: "console-config") pod "982000d1-4a44-402b-bd86-d5df1f823d5c" (UID: "982000d1-4a44-402b-bd86-d5df1f823d5c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:22:52.837237 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.837136 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "982000d1-4a44-402b-bd86-d5df1f823d5c" (UID: "982000d1-4a44-402b-bd86-d5df1f823d5c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:22:52.837309 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.837239 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-service-ca" (OuterVolumeSpecName: "service-ca") pod "982000d1-4a44-402b-bd86-d5df1f823d5c" (UID: "982000d1-4a44-402b-bd86-d5df1f823d5c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:22:52.837309 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.837263 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "982000d1-4a44-402b-bd86-d5df1f823d5c" (UID: "982000d1-4a44-402b-bd86-d5df1f823d5c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:22:52.839080 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.839043 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "982000d1-4a44-402b-bd86-d5df1f823d5c" (UID: "982000d1-4a44-402b-bd86-d5df1f823d5c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:22:52.839080 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.839047 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/982000d1-4a44-402b-bd86-d5df1f823d5c-kube-api-access-5r4xb" (OuterVolumeSpecName: "kube-api-access-5r4xb") pod "982000d1-4a44-402b-bd86-d5df1f823d5c" (UID: "982000d1-4a44-402b-bd86-d5df1f823d5c"). InnerVolumeSpecName "kube-api-access-5r4xb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 13:22:52.839222 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.839127 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "982000d1-4a44-402b-bd86-d5df1f823d5c" (UID: "982000d1-4a44-402b-bd86-d5df1f823d5c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:22:52.937744 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.937704 2579 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-oauth-serving-cert\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:22:52.937744 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.937735 2579 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-serving-cert\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:22:52.937744 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.937746 2579 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-trusted-ca-bundle\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:22:52.937744 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.937755 2579 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-5r4xb\" (UniqueName: \"kubernetes.io/projected/982000d1-4a44-402b-bd86-d5df1f823d5c-kube-api-access-5r4xb\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:22:52.938014 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.937771 2579 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-console-config\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:22:52.938014 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.937781 2579 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/982000d1-4a44-402b-bd86-d5df1f823d5c-service-ca\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:22:52.938014 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.937789 2579 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/982000d1-4a44-402b-bd86-d5df1f823d5c-console-oauth-config\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:22:52.944457 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.944419 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-54b485d8d6-8jlp9_982000d1-4a44-402b-bd86-d5df1f823d5c/console/0.log" Apr 22 13:22:52.944607 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.944481 2579 generic.go:358] "Generic (PLEG): container finished" podID="982000d1-4a44-402b-bd86-d5df1f823d5c" containerID="72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc" exitCode=2 Apr 22 13:22:52.944607 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.944522 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54b485d8d6-8jlp9" event={"ID":"982000d1-4a44-402b-bd86-d5df1f823d5c","Type":"ContainerDied","Data":"72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc"} Apr 22 13:22:52.944607 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.944544 2579 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54b485d8d6-8jlp9" Apr 22 13:22:52.944607 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.944561 2579 scope.go:117] "RemoveContainer" containerID="72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc" Apr 22 13:22:52.944754 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.944548 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54b485d8d6-8jlp9" event={"ID":"982000d1-4a44-402b-bd86-d5df1f823d5c","Type":"ContainerDied","Data":"a1117057e5e5dc1f377cc884323e7ac2c41e04ed5952b7f01fd05912f62bdb72"} Apr 22 13:22:52.952632 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.952608 2579 scope.go:117] "RemoveContainer" containerID="72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc" Apr 22 13:22:52.952911 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:22:52.952892 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc\": container with ID starting with 72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc not found: ID does not exist" containerID="72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc" Apr 22 13:22:52.952959 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.952920 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc"} err="failed to get container status \"72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc\": rpc error: code = NotFound desc = could not find container \"72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc\": container with ID starting with 72a6882a3c2cb15a45791c33d8e5a93d3250579c08be7300eef644ce794e61dc not found: ID does not exist" Apr 22 13:22:52.965059 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.965035 2579 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-54b485d8d6-8jlp9"] Apr 22 13:22:52.969592 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:52.969568 2579 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-54b485d8d6-8jlp9"] Apr 22 13:22:54.605465 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:22:54.605007 2579 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="982000d1-4a44-402b-bd86-d5df1f823d5c" path="/var/lib/kubelet/pods/982000d1-4a44-402b-bd86-d5df1f823d5c/volumes" Apr 22 13:23:00.850397 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:00.850366 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:23:00.854741 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:00.854698 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-757fc4c546-cgzdq" Apr 22 13:23:05.897135 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:05.897098 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-tmp8r" Apr 22 13:23:22.764694 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:22.764656 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:22.784670 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:22.784644 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:23.045102 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:23.045022 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:36.527301 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:36.527266 2579 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:23:36.527953 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:36.527892 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="alertmanager" containerID="cri-o://55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277" gracePeriod=120 Apr 22 13:23:36.528035 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:36.527963 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-metric" containerID="cri-o://44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56" gracePeriod=120 Apr 22 13:23:36.528035 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:36.527979 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-web" containerID="cri-o://2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a" gracePeriod=120 Apr 22 13:23:36.528136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:36.528032 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="prom-label-proxy" containerID="cri-o://196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db" gracePeriod=120 Apr 22 13:23:36.528136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:36.527992 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy" containerID="cri-o://fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3" gracePeriod=120 Apr 22 13:23:36.528136 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:36.528028 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="config-reloader" containerID="cri-o://9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f" gracePeriod=120 Apr 22 13:23:37.073311 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073278 2579 generic.go:358] "Generic (PLEG): container finished" podID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerID="196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db" exitCode=0 Apr 22 13:23:37.073311 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073304 2579 generic.go:358] "Generic (PLEG): container finished" podID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerID="fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3" exitCode=0 Apr 22 13:23:37.073311 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073312 2579 generic.go:358] "Generic (PLEG): container finished" podID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerID="9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f" exitCode=0 Apr 22 13:23:37.073311 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073319 2579 generic.go:358] "Generic (PLEG): container finished" podID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerID="55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277" exitCode=0 Apr 22 13:23:37.073606 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073355 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db"} Apr 22 13:23:37.073606 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073394 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3"} Apr 22 13:23:37.073606 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073405 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f"} Apr 22 13:23:37.073606 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.073415 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277"} Apr 22 13:23:37.777305 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.777281 2579 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:37.826579 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826547 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-tls-assets\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.826768 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826590 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-metric\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.826768 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826619 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-main-db\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.826768 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826643 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-main-tls\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.826768 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826659 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-web-config\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.826768 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826682 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-config-out\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.826768 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826708 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.826768 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826761 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.827119 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826803 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-config-volume\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.827119 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826843 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-cluster-tls-config\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.827119 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826893 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-metrics-client-ca\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.827119 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826918 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-web\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.827119 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.826945 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r67gw\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-kube-api-access-r67gw\") pod \"a758203c-d315-4a59-b93f-4fa9e34884b6\" (UID: \"a758203c-d315-4a59-b93f-4fa9e34884b6\") " Apr 22 13:23:37.828463 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.827345 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle" (OuterVolumeSpecName: "alertmanager-trusted-ca-bundle") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "alertmanager-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:23:37.828463 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.827634 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-main-db" (OuterVolumeSpecName: "alertmanager-main-db") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "alertmanager-main-db". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 13:23:37.828463 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.827998 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:23:37.830108 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.830065 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-metric" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-metric") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-metric". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:37.830463 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.830422 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 13:23:37.830605 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.830524 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-config-out" (OuterVolumeSpecName: "config-out") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 13:23:37.832082 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.832050 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-kube-api-access-r67gw" (OuterVolumeSpecName: "kube-api-access-r67gw") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "kube-api-access-r67gw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 13:23:37.832386 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.832351 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-main-tls" (OuterVolumeSpecName: "secret-alertmanager-main-tls") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "secret-alertmanager-main-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:37.832618 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.832578 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:37.832738 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.832712 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-web" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-web") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-web". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:37.832904 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.832882 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-config-volume" (OuterVolumeSpecName: "config-volume") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:37.836206 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.836176 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-cluster-tls-config" (OuterVolumeSpecName: "cluster-tls-config") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "cluster-tls-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:37.841584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.841556 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-web-config" (OuterVolumeSpecName: "web-config") pod "a758203c-d315-4a59-b93f-4fa9e34884b6" (UID: "a758203c-d315-4a59-b93f-4fa9e34884b6"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927821 2579 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-config-volume\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927853 2579 reconciler_common.go:299] "Volume detached for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-cluster-tls-config\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927864 2579 reconciler_common.go:299] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-metrics-client-ca\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927877 2579 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-web\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927887 2579 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-r67gw\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-kube-api-access-r67gw\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927897 2579 reconciler_common.go:299] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a758203c-d315-4a59-b93f-4fa9e34884b6-tls-assets\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927906 2579 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy-metric\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.927910 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927915 2579 reconciler_common.go:299] "Volume detached for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-main-db\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.928263 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927924 2579 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-main-tls\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.928263 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927933 2579 reconciler_common.go:299] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-web-config\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.928263 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927941 2579 reconciler_common.go:299] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a758203c-d315-4a59-b93f-4fa9e34884b6-config-out\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.928263 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927950 2579 reconciler_common.go:299] "Volume detached for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a758203c-d315-4a59-b93f-4fa9e34884b6-alertmanager-trusted-ca-bundle\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:37.928263 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:37.927960 2579 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a758203c-d315-4a59-b93f-4fa9e34884b6-secret-alertmanager-kube-rbac-proxy\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:38.078907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.078874 2579 generic.go:358] "Generic (PLEG): container finished" podID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerID="44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56" exitCode=0 Apr 22 13:23:38.078907 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.078903 2579 generic.go:358] "Generic (PLEG): container finished" podID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerID="2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a" exitCode=0 Apr 22 13:23:38.079114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.078946 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56"} Apr 22 13:23:38.079114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.078975 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a"} Apr 22 13:23:38.079114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.078986 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"a758203c-d315-4a59-b93f-4fa9e34884b6","Type":"ContainerDied","Data":"7f5fa518133c2c9bf9cee16f9c05a0c6f4aa65f714a3955196f0feb6564fefe2"} Apr 22 13:23:38.079114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.078988 2579 scope.go:117] "RemoveContainer" containerID="196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db" Apr 22 13:23:38.079114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.078974 2579 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.086968 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.086940 2579 scope.go:117] "RemoveContainer" containerID="44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56" Apr 22 13:23:38.094458 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.094420 2579 scope.go:117] "RemoveContainer" containerID="fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3" Apr 22 13:23:38.100766 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.100746 2579 scope.go:117] "RemoveContainer" containerID="2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a" Apr 22 13:23:38.104001 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.103980 2579 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:23:38.107681 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.107660 2579 scope.go:117] "RemoveContainer" containerID="9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f" Apr 22 13:23:38.110385 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.110360 2579 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:23:38.115440 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.115409 2579 scope.go:117] "RemoveContainer" containerID="55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277" Apr 22 13:23:38.121851 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.121825 2579 scope.go:117] "RemoveContainer" containerID="b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545" Apr 22 13:23:38.128368 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.128347 2579 scope.go:117] "RemoveContainer" containerID="196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db" Apr 22 13:23:38.128722 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:38.128700 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db\": container with ID starting with 196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db not found: ID does not exist" containerID="196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db" Apr 22 13:23:38.128794 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.128732 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db"} err="failed to get container status \"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db\": rpc error: code = NotFound desc = could not find container \"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db\": container with ID starting with 196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db not found: ID does not exist" Apr 22 13:23:38.128794 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.128753 2579 scope.go:117] "RemoveContainer" containerID="44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56" Apr 22 13:23:38.129055 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:38.129033 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56\": container with ID starting with 44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56 not found: ID does not exist" containerID="44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56" Apr 22 13:23:38.129177 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.129078 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56"} err="failed to get container status \"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56\": rpc error: code = NotFound desc = could not find container \"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56\": container with ID starting with 44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56 not found: ID does not exist" Apr 22 13:23:38.129177 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.129104 2579 scope.go:117] "RemoveContainer" containerID="fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3" Apr 22 13:23:38.129408 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:38.129389 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3\": container with ID starting with fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3 not found: ID does not exist" containerID="fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3" Apr 22 13:23:38.129495 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.129414 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3"} err="failed to get container status \"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3\": rpc error: code = NotFound desc = could not find container \"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3\": container with ID starting with fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3 not found: ID does not exist" Apr 22 13:23:38.129495 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.129451 2579 scope.go:117] "RemoveContainer" containerID="2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a" Apr 22 13:23:38.129749 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:38.129732 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a\": container with ID starting with 2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a not found: ID does not exist" containerID="2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a" Apr 22 13:23:38.129794 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.129754 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a"} err="failed to get container status \"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a\": rpc error: code = NotFound desc = could not find container \"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a\": container with ID starting with 2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a not found: ID does not exist" Apr 22 13:23:38.129832 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.129797 2579 scope.go:117] "RemoveContainer" containerID="9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f" Apr 22 13:23:38.130029 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:38.130014 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f\": container with ID starting with 9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f not found: ID does not exist" containerID="9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f" Apr 22 13:23:38.130071 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130032 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f"} err="failed to get container status \"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f\": rpc error: code = NotFound desc = could not find container \"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f\": container with ID starting with 9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f not found: ID does not exist" Apr 22 13:23:38.130071 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130045 2579 scope.go:117] "RemoveContainer" containerID="55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277" Apr 22 13:23:38.130241 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:38.130226 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277\": container with ID starting with 55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277 not found: ID does not exist" containerID="55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277" Apr 22 13:23:38.130283 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130244 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277"} err="failed to get container status \"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277\": rpc error: code = NotFound desc = could not find container \"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277\": container with ID starting with 55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277 not found: ID does not exist" Apr 22 13:23:38.130283 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130258 2579 scope.go:117] "RemoveContainer" containerID="b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545" Apr 22 13:23:38.130491 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:38.130474 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545\": container with ID starting with b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545 not found: ID does not exist" containerID="b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545" Apr 22 13:23:38.130540 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130497 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545"} err="failed to get container status \"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545\": rpc error: code = NotFound desc = could not find container \"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545\": container with ID starting with b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545 not found: ID does not exist" Apr 22 13:23:38.130540 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130514 2579 scope.go:117] "RemoveContainer" containerID="196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db" Apr 22 13:23:38.130723 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130706 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db"} err="failed to get container status \"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db\": rpc error: code = NotFound desc = could not find container \"196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db\": container with ID starting with 196edb331da771576585ea87583eb9bc45110644de6ae77e6fc1e2a84750b8db not found: ID does not exist" Apr 22 13:23:38.130766 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130723 2579 scope.go:117] "RemoveContainer" containerID="44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56" Apr 22 13:23:38.130953 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130933 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56"} err="failed to get container status \"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56\": rpc error: code = NotFound desc = could not find container \"44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56\": container with ID starting with 44cb66b06c7bbe932fb55c8736ca5577e9541807a0411a87aa09550e7b0e4e56 not found: ID does not exist" Apr 22 13:23:38.131022 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.130957 2579 scope.go:117] "RemoveContainer" containerID="fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3" Apr 22 13:23:38.131176 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131159 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3"} err="failed to get container status \"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3\": rpc error: code = NotFound desc = could not find container \"fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3\": container with ID starting with fd9745486d686696993322901c7b6ea1d9023d9bf1e6ab45720084ca0c3221c3 not found: ID does not exist" Apr 22 13:23:38.131220 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131176 2579 scope.go:117] "RemoveContainer" containerID="2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a" Apr 22 13:23:38.131394 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131376 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a"} err="failed to get container status \"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a\": rpc error: code = NotFound desc = could not find container \"2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a\": container with ID starting with 2442ab5b051bdfd6521677a7740ffd92a1028311fb8813e83b7729a3e5955b4a not found: ID does not exist" Apr 22 13:23:38.131457 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131395 2579 scope.go:117] "RemoveContainer" containerID="9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f" Apr 22 13:23:38.131687 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131672 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f"} err="failed to get container status \"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f\": rpc error: code = NotFound desc = could not find container \"9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f\": container with ID starting with 9a64ddfcafca73a7c3e488326000cd4744806d531fd7a42a8495fe9f3a1f378f not found: ID does not exist" Apr 22 13:23:38.131738 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131687 2579 scope.go:117] "RemoveContainer" containerID="55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277" Apr 22 13:23:38.131884 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131866 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277"} err="failed to get container status \"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277\": rpc error: code = NotFound desc = could not find container \"55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277\": container with ID starting with 55bf4d5a054af73d4b6aae765cb9524ae81fc79cb8daedc26b894cd01040f277 not found: ID does not exist" Apr 22 13:23:38.131925 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.131886 2579 scope.go:117] "RemoveContainer" containerID="b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545" Apr 22 13:23:38.132103 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.132086 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545"} err="failed to get container status \"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545\": rpc error: code = NotFound desc = could not find container \"b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545\": container with ID starting with b54a3a92df484d00726532057b275626e43627d61993367bade1cab83931a545 not found: ID does not exist" Apr 22 13:23:38.138980 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.138958 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:23:38.139208 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139196 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-web" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139210 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-web" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139218 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-metric" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139223 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-metric" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139230 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="alertmanager" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139235 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="alertmanager" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139243 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139248 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy" Apr 22 13:23:38.139256 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139255 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="config-reloader" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139261 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="config-reloader" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139273 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="prom-label-proxy" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139278 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="prom-label-proxy" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139286 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="init-config-reloader" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139291 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="init-config-reloader" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139300 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="982000d1-4a44-402b-bd86-d5df1f823d5c" containerName="console" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139305 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="982000d1-4a44-402b-bd86-d5df1f823d5c" containerName="console" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139345 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="alertmanager" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139352 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-metric" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139358 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="config-reloader" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139364 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139378 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="982000d1-4a44-402b-bd86-d5df1f823d5c" containerName="console" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139385 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="prom-label-proxy" Apr 22 13:23:38.139513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.139392 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" containerName="kube-rbac-proxy-web" Apr 22 13:23:38.144210 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.144191 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.146961 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.146939 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Apr 22 13:23:38.147059 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.146966 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Apr 22 13:23:38.147059 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.146939 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Apr 22 13:23:38.147284 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.147266 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-r4gw2\"" Apr 22 13:23:38.147365 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.147291 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Apr 22 13:23:38.147365 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.147272 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Apr 22 13:23:38.147365 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.147320 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Apr 22 13:23:38.147560 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.147543 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Apr 22 13:23:38.147663 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.147650 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Apr 22 13:23:38.153103 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.153060 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Apr 22 13:23:38.157128 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.157101 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:23:38.229573 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229477 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/443b8094-21ec-4417-9c5d-defe0b0a1326-config-out\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229573 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229522 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229573 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229545 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/443b8094-21ec-4417-9c5d-defe0b0a1326-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229789 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229622 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-web-config\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229789 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229659 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229789 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229701 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/443b8094-21ec-4417-9c5d-defe0b0a1326-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229789 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229745 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/443b8094-21ec-4417-9c5d-defe0b0a1326-tls-assets\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229789 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229773 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229806 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/443b8094-21ec-4417-9c5d-defe0b0a1326-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229827 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-config-volume\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229847 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229899 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.229934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.229921 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntgzv\" (UniqueName: \"kubernetes.io/projected/443b8094-21ec-4417-9c5d-defe0b0a1326-kube-api-access-ntgzv\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330581 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330546 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330591 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330607 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ntgzv\" (UniqueName: \"kubernetes.io/projected/443b8094-21ec-4417-9c5d-defe0b0a1326-kube-api-access-ntgzv\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330631 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/443b8094-21ec-4417-9c5d-defe0b0a1326-config-out\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330654 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330671 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/443b8094-21ec-4417-9c5d-defe0b0a1326-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330701 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-web-config\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.330737 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330724 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.331117 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330773 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/443b8094-21ec-4417-9c5d-defe0b0a1326-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.331117 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330808 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/443b8094-21ec-4417-9c5d-defe0b0a1326-tls-assets\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.331117 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330840 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.331117 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330873 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/443b8094-21ec-4417-9c5d-defe0b0a1326-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.331117 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.330899 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-config-volume\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.331365 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.331250 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/443b8094-21ec-4417-9c5d-defe0b0a1326-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.331965 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.331910 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/443b8094-21ec-4417-9c5d-defe0b0a1326-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.332526 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.332501 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/443b8094-21ec-4417-9c5d-defe0b0a1326-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.333885 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.333860 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.334003 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.333982 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.334338 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.334314 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/443b8094-21ec-4417-9c5d-defe0b0a1326-config-out\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.334420 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.334406 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-web-config\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.334509 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.334487 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.334550 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.334525 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/443b8094-21ec-4417-9c5d-defe0b0a1326-tls-assets\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.334598 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.334583 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.334639 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.334626 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-config-volume\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.335391 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.335367 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/443b8094-21ec-4417-9c5d-defe0b0a1326-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.340135 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.340117 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntgzv\" (UniqueName: \"kubernetes.io/projected/443b8094-21ec-4417-9c5d-defe0b0a1326-kube-api-access-ntgzv\") pod \"alertmanager-main-0\" (UID: \"443b8094-21ec-4417-9c5d-defe0b0a1326\") " pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.453592 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.453539 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Apr 22 13:23:38.580791 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.580745 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Apr 22 13:23:38.584670 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:23:38.584643 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod443b8094_21ec_4417_9c5d_defe0b0a1326.slice/crio-05b4f662e5fa3c317cb8a7f1a42dfc3f43e3375c2ddcbc52a1dc613cbf05bdb2 WatchSource:0}: Error finding container 05b4f662e5fa3c317cb8a7f1a42dfc3f43e3375c2ddcbc52a1dc613cbf05bdb2: Status 404 returned error can't find the container with id 05b4f662e5fa3c317cb8a7f1a42dfc3f43e3375c2ddcbc52a1dc613cbf05bdb2 Apr 22 13:23:38.608713 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:38.608654 2579 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a758203c-d315-4a59-b93f-4fa9e34884b6" path="/var/lib/kubelet/pods/a758203c-d315-4a59-b93f-4fa9e34884b6/volumes" Apr 22 13:23:39.082969 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:39.082932 2579 generic.go:358] "Generic (PLEG): container finished" podID="443b8094-21ec-4417-9c5d-defe0b0a1326" containerID="9a4660298769ccc80a0670e6b4f1dba53c13ad1cf5fedf155c9fb1e6a1913f4b" exitCode=0 Apr 22 13:23:39.083455 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:39.083021 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerDied","Data":"9a4660298769ccc80a0670e6b4f1dba53c13ad1cf5fedf155c9fb1e6a1913f4b"} Apr 22 13:23:39.083455 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:39.083059 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerStarted","Data":"05b4f662e5fa3c317cb8a7f1a42dfc3f43e3375c2ddcbc52a1dc613cbf05bdb2"} Apr 22 13:23:40.090410 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.090375 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerStarted","Data":"4c1d9ab3e002525e007ecd2e8c962b438e7730e2c1426242280541111680db94"} Apr 22 13:23:40.090410 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.090413 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerStarted","Data":"763d860267f42a8487aac4ddfbbddc9e66853c49f705e2ee258b805a2e24bb65"} Apr 22 13:23:40.090834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.090423 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerStarted","Data":"719624e11127ef7ddec30d250ceedfff49797e0439298cec0c542d316896fd0b"} Apr 22 13:23:40.090834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.090454 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerStarted","Data":"e71a83fd648ca5262e464eab41e0695e5f3f71297aeb77885880cde30fcc9a4f"} Apr 22 13:23:40.090834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.090463 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerStarted","Data":"e16a96bd3d8d85717eeda90734df6a34eeeb90883199ed2a69ac65747867ca43"} Apr 22 13:23:40.090834 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.090471 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"443b8094-21ec-4417-9c5d-defe0b0a1326","Type":"ContainerStarted","Data":"66971d538cb2cba3fb2e734a120d9321c3634906d6759930b63d06e146543c8a"} Apr 22 13:23:40.121153 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.120479 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.120453326 podStartE2EDuration="2.120453326s" podCreationTimestamp="2026-04-22 13:23:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 13:23:40.11836777 +0000 UTC m=+138.054754659" watchObservedRunningTime="2026-04-22 13:23:40.120453326 +0000 UTC m=+138.056840186" Apr 22 13:23:40.527567 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.527530 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/telemeter-client-6cd4f57488-jvpmt"] Apr 22 13:23:40.530786 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.530763 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.534113 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.534092 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-kube-rbac-proxy-config\"" Apr 22 13:23:40.534225 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.534092 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemeter-client-serving-certs-ca-bundle\"" Apr 22 13:23:40.534225 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.534188 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client\"" Apr 22 13:23:40.534584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.534188 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-tls\"" Apr 22 13:23:40.534998 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.534945 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"telemeter-client-dockercfg-5d959\"" Apr 22 13:23:40.535154 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.535075 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"federate-client-certs\"" Apr 22 13:23:40.541413 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.541390 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemeter-trusted-ca-bundle-8i12ta5c71j38\"" Apr 22 13:23:40.544560 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.544535 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-6cd4f57488-jvpmt"] Apr 22 13:23:40.545557 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545537 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-serving-certs-ca-bundle\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.545657 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545577 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-federate-client-tls\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.545657 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545613 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-secret-telemeter-client\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.545769 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545654 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-metrics-client-ca\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.545769 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545713 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9llxr\" (UniqueName: \"kubernetes.io/projected/00f81ad6-66db-48b9-9348-edc079cfcd69-kube-api-access-9llxr\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.545769 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545742 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.545880 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545786 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-telemeter-client-tls\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.545880 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.545851 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-telemeter-trusted-ca-bundle\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647109 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647056 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-metrics-client-ca\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647109 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647117 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9llxr\" (UniqueName: \"kubernetes.io/projected/00f81ad6-66db-48b9-9348-edc079cfcd69-kube-api-access-9llxr\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647140 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647170 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-telemeter-client-tls\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647191 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-telemeter-trusted-ca-bundle\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647234 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-serving-certs-ca-bundle\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647281 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-federate-client-tls\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.647384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.647319 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-secret-telemeter-client\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.648106 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.648076 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-serving-certs-ca-bundle\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.648250 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.648164 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-metrics-client-ca\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.648506 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.648486 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00f81ad6-66db-48b9-9348-edc079cfcd69-telemeter-trusted-ca-bundle\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.649869 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.649845 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.650011 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.649988 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-telemeter-client-tls\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.650218 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.650204 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-federate-client-tls\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.650251 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.650233 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/00f81ad6-66db-48b9-9348-edc079cfcd69-secret-telemeter-client\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.655270 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.655244 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9llxr\" (UniqueName: \"kubernetes.io/projected/00f81ad6-66db-48b9-9348-edc079cfcd69-kube-api-access-9llxr\") pod \"telemeter-client-6cd4f57488-jvpmt\" (UID: \"00f81ad6-66db-48b9-9348-edc079cfcd69\") " pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.801584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.801496 2579 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:23:40.802112 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.801976 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy" containerID="cri-o://2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66" gracePeriod=600 Apr 22 13:23:40.802112 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.801999 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="config-reloader" containerID="cri-o://ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb" gracePeriod=600 Apr 22 13:23:40.802112 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.802018 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-web" containerID="cri-o://0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554" gracePeriod=600 Apr 22 13:23:40.802112 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.802048 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-thanos" containerID="cri-o://3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751" gracePeriod=600 Apr 22 13:23:40.802416 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.801990 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="thanos-sidecar" containerID="cri-o://2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e" gracePeriod=600 Apr 22 13:23:40.802416 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.801967 2579 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="prometheus" containerID="cri-o://2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8" gracePeriod=600 Apr 22 13:23:40.845025 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.844998 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" Apr 22 13:23:40.972104 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:40.972073 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-6cd4f57488-jvpmt"] Apr 22 13:23:40.974191 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:23:40.974161 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00f81ad6_66db_48b9_9348_edc079cfcd69.slice/crio-110391529e7b09676f13d2dc1537e945f019a2fdf02af9a591b685494774d1b8 WatchSource:0}: Error finding container 110391529e7b09676f13d2dc1537e945f019a2fdf02af9a591b685494774d1b8: Status 404 returned error can't find the container with id 110391529e7b09676f13d2dc1537e945f019a2fdf02af9a591b685494774d1b8 Apr 22 13:23:41.095384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.095291 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" event={"ID":"00f81ad6-66db-48b9-9348-edc079cfcd69","Type":"ContainerStarted","Data":"110391529e7b09676f13d2dc1537e945f019a2fdf02af9a591b685494774d1b8"} Apr 22 13:23:41.098171 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098144 2579 generic.go:358] "Generic (PLEG): container finished" podID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerID="3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751" exitCode=0 Apr 22 13:23:41.098171 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098169 2579 generic.go:358] "Generic (PLEG): container finished" podID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerID="2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66" exitCode=0 Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098179 2579 generic.go:358] "Generic (PLEG): container finished" podID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerID="2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e" exitCode=0 Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098187 2579 generic.go:358] "Generic (PLEG): container finished" podID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerID="ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb" exitCode=0 Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098195 2579 generic.go:358] "Generic (PLEG): container finished" podID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerID="2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8" exitCode=0 Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098208 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751"} Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098237 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66"} Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098248 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e"} Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098257 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb"} Apr 22 13:23:41.098382 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:41.098267 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8"} Apr 22 13:23:42.055479 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.055453 2579 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.104828 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.104799 2579 generic.go:358] "Generic (PLEG): container finished" podID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerID="0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554" exitCode=0 Apr 22 13:23:42.105250 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.104875 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554"} Apr 22 13:23:42.105250 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.104923 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7","Type":"ContainerDied","Data":"2da4f9c3ab245741675f2b1ebb0e5aeddb654c8bbd4f6a041a6348dc191ae6b1"} Apr 22 13:23:42.105250 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.104940 2579 scope.go:117] "RemoveContainer" containerID="3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751" Apr 22 13:23:42.105250 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.104941 2579 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.112748 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.112722 2579 scope.go:117] "RemoveContainer" containerID="2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66" Apr 22 13:23:42.120140 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.120114 2579 scope.go:117] "RemoveContainer" containerID="0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554" Apr 22 13:23:42.127580 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.127559 2579 scope.go:117] "RemoveContainer" containerID="2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e" Apr 22 13:23:42.136514 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.136490 2579 scope.go:117] "RemoveContainer" containerID="ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb" Apr 22 13:23:42.144253 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.144231 2579 scope.go:117] "RemoveContainer" containerID="2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8" Apr 22 13:23:42.153075 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.153055 2579 scope.go:117] "RemoveContainer" containerID="6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8" Apr 22 13:23:42.158237 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158213 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158360 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158245 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-tls\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158360 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158270 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-trusted-ca-bundle\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158360 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158300 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-tls-assets\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158360 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158331 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9565\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-kube-api-access-t9565\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158593 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158368 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-metrics-client-ca\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158593 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158391 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-grpc-tls\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158593 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158417 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-thanos-prometheus-http-client-file\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158593 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158470 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-metrics-client-certs\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158593 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158510 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-kubelet-serving-ca-bundle\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158593 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158539 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-serving-certs-ca-bundle\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158593 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158570 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-kube-rbac-proxy\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158920 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158610 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158920 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158670 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-rulefiles-0\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.158920 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.158700 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config-out\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.159690 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-db\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.159761 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-web-config\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.159792 2579 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\" (UID: \"8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7\") " Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.159858 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-trusted-ca-bundle" (OuterVolumeSpecName: "prometheus-trusted-ca-bundle") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "prometheus-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.160095 2579 reconciler_common.go:299] "Volume detached for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-trusted-ca-bundle\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.161105 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-kubelet-serving-ca-bundle" (OuterVolumeSpecName: "configmap-kubelet-serving-ca-bundle") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "configmap-kubelet-serving-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.161275 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-rulefiles-0" (OuterVolumeSpecName: "prometheus-k8s-rulefiles-0") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "prometheus-k8s-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.161364 2579 scope.go:117] "RemoveContainer" containerID="3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.161378 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-serving-certs-ca-bundle" (OuterVolumeSpecName: "configmap-serving-certs-ca-bundle") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "configmap-serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.161862 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-metrics-client-ca" (OuterVolumeSpecName: "configmap-metrics-client-ca") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "configmap-metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:42.161953 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751\": container with ID starting with 3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751 not found: ID does not exist" containerID="3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.161986 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751"} err="failed to get container status \"3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751\": rpc error: code = NotFound desc = could not find container \"3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751\": container with ID starting with 3152f6f5066cf200ca82fd8755812ba3eddab3c695ec3e89c49f021caca8e751 not found: ID does not exist" Apr 22 13:23:42.162322 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.162012 2579 scope.go:117] "RemoveContainer" containerID="2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66" Apr 22 13:23:42.163114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.162468 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-db" (OuterVolumeSpecName: "prometheus-k8s-db") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "prometheus-k8s-db". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 13:23:42.163114 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:42.162541 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66\": container with ID starting with 2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66 not found: ID does not exist" containerID="2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66" Apr 22 13:23:42.163114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.162569 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66"} err="failed to get container status \"2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66\": rpc error: code = NotFound desc = could not find container \"2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66\": container with ID starting with 2b5f6b583b4410a7e2558dcc083de374ab3619c64eeaf393cda2bdcc976b6f66 not found: ID does not exist" Apr 22 13:23:42.163114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.162591 2579 scope.go:117] "RemoveContainer" containerID="0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554" Apr 22 13:23:42.163114 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.162968 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-tls") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "secret-prometheus-k8s-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.163114 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:42.163089 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554\": container with ID starting with 0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554 not found: ID does not exist" containerID="0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554" Apr 22 13:23:42.163569 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.163115 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554"} err="failed to get container status \"0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554\": rpc error: code = NotFound desc = could not find container \"0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554\": container with ID starting with 0ec845ab7077b2a6bd5d7ab2f503bdbc6931393476c50ea483acff486c619554 not found: ID does not exist" Apr 22 13:23:42.163569 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.163135 2579 scope.go:117] "RemoveContainer" containerID="2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e" Apr 22 13:23:42.163569 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:42.163397 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e\": container with ID starting with 2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e not found: ID does not exist" containerID="2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e" Apr 22 13:23:42.163569 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.163450 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e"} err="failed to get container status \"2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e\": rpc error: code = NotFound desc = could not find container \"2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e\": container with ID starting with 2ba99139488a8944f2cf127da876430955f61adcc1bb0d50bbd72aebcd17f66e not found: ID does not exist" Apr 22 13:23:42.163569 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.163471 2579 scope.go:117] "RemoveContainer" containerID="ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb" Apr 22 13:23:42.163861 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:42.163685 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb\": container with ID starting with ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb not found: ID does not exist" containerID="ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb" Apr 22 13:23:42.163861 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.163711 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb"} err="failed to get container status \"ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb\": rpc error: code = NotFound desc = could not find container \"ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb\": container with ID starting with ca66f54a010a217134b7e7d277be9ff947af32c7e208e0e28c46f202df25a4fb not found: ID does not exist" Apr 22 13:23:42.163861 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.163731 2579 scope.go:117] "RemoveContainer" containerID="2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8" Apr 22 13:23:42.164076 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:42.164049 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8\": container with ID starting with 2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8 not found: ID does not exist" containerID="2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8" Apr 22 13:23:42.164155 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164080 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8"} err="failed to get container status \"2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8\": rpc error: code = NotFound desc = could not find container \"2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8\": container with ID starting with 2b7f7a503940bb8d17ff9354a59cb191d24334d567ab653742ffb439ccaa72c8 not found: ID does not exist" Apr 22 13:23:42.164155 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164101 2579 scope.go:117] "RemoveContainer" containerID="6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8" Apr 22 13:23:42.164605 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164363 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-kube-rbac-proxy-web" (OuterVolumeSpecName: "secret-prometheus-k8s-kube-rbac-proxy-web") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "secret-prometheus-k8s-kube-rbac-proxy-web". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.164605 ip-10-0-128-54 kubenswrapper[2579]: E0422 13:23:42.164503 2579 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8\": container with ID starting with 6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8 not found: ID does not exist" containerID="6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8" Apr 22 13:23:42.164605 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164531 2579 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8"} err="failed to get container status \"6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8\": rpc error: code = NotFound desc = could not find container \"6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8\": container with ID starting with 6517789ef69151a61fc5e12eca0e82c96c5237f2fe875c92de0675c4fdd631a8 not found: ID does not exist" Apr 22 13:23:42.164811 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164745 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config-out" (OuterVolumeSpecName: "config-out") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 13:23:42.164811 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164753 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config" (OuterVolumeSpecName: "config") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.164913 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164891 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-kube-api-access-t9565" (OuterVolumeSpecName: "kube-api-access-t9565") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "kube-api-access-t9565". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 13:23:42.164964 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.164942 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-thanos-sidecar-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-thanos-sidecar-tls") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "secret-prometheus-k8s-thanos-sidecar-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.165072 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.165051 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-metrics-client-certs" (OuterVolumeSpecName: "secret-metrics-client-certs") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "secret-metrics-client-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.165186 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.165137 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 13:23:42.165257 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.165188 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.165315 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.165255 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-grpc-tls" (OuterVolumeSpecName: "secret-grpc-tls") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "secret-grpc-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.165315 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.165298 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-kube-rbac-proxy" (OuterVolumeSpecName: "secret-kube-rbac-proxy") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "secret-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.177200 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.177173 2579 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-web-config" (OuterVolumeSpecName: "web-config") pod "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" (UID: "8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 13:23:42.260473 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260407 2579 reconciler_common.go:299] "Volume detached for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-metrics-client-certs\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260473 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260475 2579 reconciler_common.go:299] "Volume detached for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-kubelet-serving-ca-bundle\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260486 2579 reconciler_common.go:299] "Volume detached for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-serving-certs-ca-bundle\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260496 2579 reconciler_common.go:299] "Volume detached for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-kube-rbac-proxy\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260507 2579 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-thanos-sidecar-tls\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260518 2579 reconciler_common.go:299] "Volume detached for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-rulefiles-0\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260529 2579 reconciler_common.go:299] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config-out\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260538 2579 reconciler_common.go:299] "Volume detached for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-prometheus-k8s-db\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260548 2579 reconciler_common.go:299] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-web-config\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260557 2579 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-kube-rbac-proxy-web\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260565 2579 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-config\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260576 2579 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-prometheus-k8s-tls\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260587 2579 reconciler_common.go:299] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-tls-assets\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260595 2579 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-t9565\" (UniqueName: \"kubernetes.io/projected/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-kube-api-access-t9565\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260604 2579 reconciler_common.go:299] "Volume detached for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-configmap-metrics-client-ca\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260612 2579 reconciler_common.go:299] "Volume detached for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-secret-grpc-tls\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.260654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.260622 2579 reconciler_common.go:299] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7-thanos-prometheus-http-client-file\") on node \"ip-10-0-128-54.ec2.internal\" DevicePath \"\"" Apr 22 13:23:42.435192 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.435150 2579 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:23:42.438779 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.438744 2579 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:23:42.467968 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.467934 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:23:42.468376 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468355 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="prometheus" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468394 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="prometheus" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468413 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-web" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468422 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-web" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468456 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-thanos" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468465 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-thanos" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468473 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468483 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468496 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="init-config-reloader" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468507 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="init-config-reloader" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468515 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="thanos-sidecar" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468522 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="thanos-sidecar" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468537 2579 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="config-reloader" Apr 22 13:23:42.468549 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468544 2579 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="config-reloader" Apr 22 13:23:42.469129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468626 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-web" Apr 22 13:23:42.469129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468637 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy" Apr 22 13:23:42.469129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468651 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="config-reloader" Apr 22 13:23:42.469129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468660 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="thanos-sidecar" Apr 22 13:23:42.469129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468671 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="prometheus" Apr 22 13:23:42.469129 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.468681 2579 memory_manager.go:356] "RemoveStaleState removing state" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" containerName="kube-rbac-proxy-thanos" Apr 22 13:23:42.477896 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.474808 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.484700 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.484673 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Apr 22 13:23:42.485117 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485098 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Apr 22 13:23:42.494862 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485561 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Apr 22 13:23:42.494998 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485609 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-2dq2oeagafuup\"" Apr 22 13:23:42.495070 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485679 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Apr 22 13:23:42.495123 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485738 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Apr 22 13:23:42.495180 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485885 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Apr 22 13:23:42.495238 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485930 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Apr 22 13:23:42.495287 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.485999 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Apr 22 13:23:42.495337 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.486070 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Apr 22 13:23:42.495337 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.486261 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Apr 22 13:23:42.495464 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.486306 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-llln8\"" Apr 22 13:23:42.495464 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.488263 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Apr 22 13:23:42.495464 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.490716 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:23:42.495464 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.492053 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Apr 22 13:23:42.563513 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563479 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563691 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563525 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563691 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563549 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563691 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563601 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-config\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563691 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563627 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563691 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563655 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/148bd9c7-9015-47c6-9ee3-1fcb791544a6-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563714 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95xrm\" (UniqueName: \"kubernetes.io/projected/148bd9c7-9015-47c6-9ee3-1fcb791544a6-kube-api-access-95xrm\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563769 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563800 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563836 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/148bd9c7-9015-47c6-9ee3-1fcb791544a6-config-out\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563869 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.563934 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563897 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.564141 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.563971 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.564141 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.564011 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.564141 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.564048 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.564141 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.564074 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.564141 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.564098 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-web-config\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.564141 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.564139 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.605307 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.605272 2579 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7" path="/var/lib/kubelet/pods/8d9eec3c-17ba-4fcd-ac61-5fdbc1dc7ed7/volumes" Apr 22 13:23:42.664654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664573 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664611 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664654 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664642 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-config\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664659 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664679 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/148bd9c7-9015-47c6-9ee3-1fcb791544a6-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664700 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-95xrm\" (UniqueName: \"kubernetes.io/projected/148bd9c7-9015-47c6-9ee3-1fcb791544a6-kube-api-access-95xrm\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664731 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664759 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664791 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/148bd9c7-9015-47c6-9ee3-1fcb791544a6-config-out\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664824 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664848 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.664894 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664880 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665236 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664906 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665236 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664939 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665236 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664963 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665236 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.664984 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-web-config\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665236 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.665024 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665236 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.665070 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665236 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.665157 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.665847 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.665818 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.666397 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.666176 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.667357 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.667124 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.668723 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.668700 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.669377 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.668751 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148bd9c7-9015-47c6-9ee3-1fcb791544a6-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.669377 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.669062 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/148bd9c7-9015-47c6-9ee3-1fcb791544a6-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.669377 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.669286 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.670155 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.670108 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.671009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.670821 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-web-config\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.671009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.670900 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.671009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.670953 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.671009 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.670955 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.671409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.671382 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-config\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.671498 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.671395 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/148bd9c7-9015-47c6-9ee3-1fcb791544a6-config-out\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.671498 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.671472 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.672465 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.672409 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/148bd9c7-9015-47c6-9ee3-1fcb791544a6-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.675352 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.675332 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-95xrm\" (UniqueName: \"kubernetes.io/projected/148bd9c7-9015-47c6-9ee3-1fcb791544a6-kube-api-access-95xrm\") pod \"prometheus-k8s-0\" (UID: \"148bd9c7-9015-47c6-9ee3-1fcb791544a6\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:42.803590 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:42.803548 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:23:43.229229 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:43.229208 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 22 13:23:43.231544 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:23:43.231517 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod148bd9c7_9015_47c6_9ee3_1fcb791544a6.slice/crio-3320d78b0114690c179e357bf88501ea4b6868700eb7f20105710aa513f8f627 WatchSource:0}: Error finding container 3320d78b0114690c179e357bf88501ea4b6868700eb7f20105710aa513f8f627: Status 404 returned error can't find the container with id 3320d78b0114690c179e357bf88501ea4b6868700eb7f20105710aa513f8f627 Apr 22 13:23:44.115917 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:44.115875 2579 generic.go:358] "Generic (PLEG): container finished" podID="148bd9c7-9015-47c6-9ee3-1fcb791544a6" containerID="d01eb16bb5d6e8732fde84e10d7b45f0c415cc208b2c1ed9eaa609db82192769" exitCode=0 Apr 22 13:23:44.116099 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:44.115957 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerDied","Data":"d01eb16bb5d6e8732fde84e10d7b45f0c415cc208b2c1ed9eaa609db82192769"} Apr 22 13:23:44.116099 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:44.115993 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerStarted","Data":"3320d78b0114690c179e357bf88501ea4b6868700eb7f20105710aa513f8f627"} Apr 22 13:23:44.117926 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:44.117900 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" event={"ID":"00f81ad6-66db-48b9-9348-edc079cfcd69","Type":"ContainerStarted","Data":"395d50128696e0c449304be19e9c32a8d4daad1176859278d5d6aa1f81dca2be"} Apr 22 13:23:44.118031 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:44.117929 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" event={"ID":"00f81ad6-66db-48b9-9348-edc079cfcd69","Type":"ContainerStarted","Data":"c8e1f8a9170f7626f7fe9ced7338f02d11204ab865694bfb8dc58c8cb73ffba6"} Apr 22 13:23:44.118031 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:44.117940 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" event={"ID":"00f81ad6-66db-48b9-9348-edc079cfcd69","Type":"ContainerStarted","Data":"b514305714c4dbb7ba6391389b3f6cf2cba2113c4ef12d5c75b6491ed2fdf5f6"} Apr 22 13:23:44.170804 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:44.170760 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/telemeter-client-6cd4f57488-jvpmt" podStartSLOduration=2.006245909 podStartE2EDuration="4.170744976s" podCreationTimestamp="2026-04-22 13:23:40 +0000 UTC" firstStartedPulling="2026-04-22 13:23:40.975974801 +0000 UTC m=+138.912361645" lastFinishedPulling="2026-04-22 13:23:43.140473852 +0000 UTC m=+141.076860712" observedRunningTime="2026-04-22 13:23:44.169066784 +0000 UTC m=+142.105453659" watchObservedRunningTime="2026-04-22 13:23:44.170744976 +0000 UTC m=+142.107131873" Apr 22 13:23:45.128789 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:45.128752 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerStarted","Data":"1fd47f6ce91facfe3a9dad43400848167946fa11417ffdadb21e670a9e5b971e"} Apr 22 13:23:45.128789 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:45.128794 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerStarted","Data":"11c745db07539a78e8345c0659f6c5be1258f1924d188e02be37224af00abb6c"} Apr 22 13:23:45.129313 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:45.128806 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerStarted","Data":"314fbeb0ce0f7ece9d05895d354be59961abbd299f28df618747ccecc2aa169f"} Apr 22 13:23:45.129313 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:45.128820 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerStarted","Data":"d7120e845170855e5bca08fc343f7a97e60b30a17535986c8821450bf73181d8"} Apr 22 13:23:45.129313 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:45.128829 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerStarted","Data":"d2c99d0018b6ad2ff4709d267ce21397f16a1e15225f2d2648705e8754704379"} Apr 22 13:23:45.129313 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:45.128837 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"148bd9c7-9015-47c6-9ee3-1fcb791544a6","Type":"ContainerStarted","Data":"483deef49c8e9f922891e80870dc4e8e16cf96ec352d13472e261b7b212b7b4a"} Apr 22 13:23:45.160957 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:45.160896 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=3.160876836 podStartE2EDuration="3.160876836s" podCreationTimestamp="2026-04-22 13:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 13:23:45.158850603 +0000 UTC m=+143.095237470" watchObservedRunningTime="2026-04-22 13:23:45.160876836 +0000 UTC m=+143.097263701" Apr 22 13:23:47.804292 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:23:47.804248 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:24:42.804332 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:24:42.804290 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:24:42.819441 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:24:42.819400 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:24:43.305638 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:24:43.305608 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 22 13:26:22.487100 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:26:22.487069 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:26:22.487649 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:26:22.487253 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:26:22.493925 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:26:22.493904 2579 kubelet.go:1628] "Image garbage collection succeeded" Apr 22 13:27:40.813388 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.813355 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp"] Apr 22 13:27:40.816495 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.816473 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.819062 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.819018 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-jobset-operator\"/\"kube-root-ca.crt\"" Apr 22 13:27:40.819062 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.819035 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-jobset-operator\"/\"openshift-service-ca.crt\"" Apr 22 13:27:40.819230 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.819073 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-jobset-operator\"/\"webhook-server-cert\"" Apr 22 13:27:40.819230 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.819035 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-jobset-operator\"/\"jobset-controller-manager-dockercfg-q9g4b\"" Apr 22 13:27:40.819230 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.819018 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-jobset-operator\"/\"jobset-manager-config\"" Apr 22 13:27:40.820040 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.820024 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-jobset-operator\"/\"metrics-server-cert\"" Apr 22 13:27:40.826997 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.826977 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp"] Apr 22 13:27:40.881152 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.881126 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhrmr\" (UniqueName: \"kubernetes.io/projected/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-kube-api-access-jhrmr\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.881263 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.881189 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-cert\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.881263 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.881230 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-metrics-certs\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.881345 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.881312 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-manager-config\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.981756 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.981735 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-manager-config\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.981829 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.981779 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jhrmr\" (UniqueName: \"kubernetes.io/projected/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-kube-api-access-jhrmr\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.981829 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.981816 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-cert\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.981930 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.981835 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-metrics-certs\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.982335 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.982314 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-manager-config\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.984197 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.984175 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-metrics-certs\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.984279 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.984252 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-cert\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:40.994782 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:40.994757 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhrmr\" (UniqueName: \"kubernetes.io/projected/f4bca08f-b935-4e6c-8f5b-3e16f5e93352-kube-api-access-jhrmr\") pod \"jobset-controller-manager-7876c9c48-fcxtp\" (UID: \"f4bca08f-b935-4e6c-8f5b-3e16f5e93352\") " pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:41.126289 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:41.126220 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:41.240035 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:41.240006 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp"] Apr 22 13:27:41.243261 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:27:41.243235 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4bca08f_b935_4e6c_8f5b_3e16f5e93352.slice/crio-276fae0268d7e32d9adf7ede3398af340fddd6203ec51e22f3455f121f016ebf WatchSource:0}: Error finding container 276fae0268d7e32d9adf7ede3398af340fddd6203ec51e22f3455f121f016ebf: Status 404 returned error can't find the container with id 276fae0268d7e32d9adf7ede3398af340fddd6203ec51e22f3455f121f016ebf Apr 22 13:27:41.244966 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:41.244944 2579 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 13:27:41.765974 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:41.765936 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" event={"ID":"f4bca08f-b935-4e6c-8f5b-3e16f5e93352","Type":"ContainerStarted","Data":"276fae0268d7e32d9adf7ede3398af340fddd6203ec51e22f3455f121f016ebf"} Apr 22 13:27:44.775999 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:44.775962 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" event={"ID":"f4bca08f-b935-4e6c-8f5b-3e16f5e93352","Type":"ContainerStarted","Data":"377abaab4720eea35c34712e974ec3093dcf9564b53cf5cd17920a0cec66419b"} Apr 22 13:27:44.776344 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:44.776157 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:27:44.794367 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:44.794329 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" podStartSLOduration=1.882386393 podStartE2EDuration="4.794318356s" podCreationTimestamp="2026-04-22 13:27:40 +0000 UTC" firstStartedPulling="2026-04-22 13:27:41.245066157 +0000 UTC m=+379.181452998" lastFinishedPulling="2026-04-22 13:27:44.156998108 +0000 UTC m=+382.093384961" observedRunningTime="2026-04-22 13:27:44.792872063 +0000 UTC m=+382.729258925" watchObservedRunningTime="2026-04-22 13:27:44.794318356 +0000 UTC m=+382.730705217" Apr 22 13:27:55.783859 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:27:55.783831 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-jobset-operator/jobset-controller-manager-7876c9c48-fcxtp" Apr 22 13:31:22.506279 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:31:22.506251 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:31:22.508190 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:31:22.508168 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:36:22.529835 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:36:22.529746 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:36:22.532517 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:36:22.532496 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:39:07.152469 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.152380 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-6b79c7bb88-p8w9c"] Apr 22 13:39:07.155402 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.155384 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.158027 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.158002 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 22 13:39:07.159102 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.159084 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-rgk2k\"" Apr 22 13:39:07.159201 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.159122 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 22 13:39:07.159201 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.159138 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 22 13:39:07.159201 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.159127 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 22 13:39:07.159371 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.159126 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 22 13:39:07.159371 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.159127 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 22 13:39:07.159515 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.159499 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 22 13:39:07.162905 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.162885 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 22 13:39:07.167409 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.167390 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b79c7bb88-p8w9c"] Apr 22 13:39:07.220738 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.220720 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-serving-cert\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.220833 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.220747 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngdwt\" (UniqueName: \"kubernetes.io/projected/615924fd-d8e9-48d2-8c72-6a4cb484cc10-kube-api-access-ngdwt\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.220833 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.220774 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-config\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.220833 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.220813 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-oauth-serving-cert\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.220927 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.220843 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-service-ca\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.220927 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.220894 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-oauth-config\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.220927 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.220916 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-trusted-ca-bundle\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.321757 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.321737 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-oauth-config\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.321899 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.321769 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-trusted-ca-bundle\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.321899 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.321789 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-serving-cert\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.321899 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.321807 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ngdwt\" (UniqueName: \"kubernetes.io/projected/615924fd-d8e9-48d2-8c72-6a4cb484cc10-kube-api-access-ngdwt\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.321899 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.321892 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-config\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.322074 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.321916 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-oauth-serving-cert\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.322074 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.321939 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-service-ca\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.322611 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.322590 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-service-ca\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.322678 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.322593 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-oauth-serving-cert\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.322678 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.322594 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-config\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.322678 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.322650 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/615924fd-d8e9-48d2-8c72-6a4cb484cc10-trusted-ca-bundle\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.324621 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.324597 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-oauth-config\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.324621 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.324610 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/615924fd-d8e9-48d2-8c72-6a4cb484cc10-console-serving-cert\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.329584 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.329564 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngdwt\" (UniqueName: \"kubernetes.io/projected/615924fd-d8e9-48d2-8c72-6a4cb484cc10-kube-api-access-ngdwt\") pod \"console-6b79c7bb88-p8w9c\" (UID: \"615924fd-d8e9-48d2-8c72-6a4cb484cc10\") " pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.464620 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.464567 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:07.582908 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.582859 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b79c7bb88-p8w9c"] Apr 22 13:39:07.585720 ip-10-0-128-54 kubenswrapper[2579]: W0422 13:39:07.585691 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod615924fd_d8e9_48d2_8c72_6a4cb484cc10.slice/crio-b370c88452268c3d7cb54dcf12872afd1ae627c89e61fa519ecd42fc93ae05d2 WatchSource:0}: Error finding container b370c88452268c3d7cb54dcf12872afd1ae627c89e61fa519ecd42fc93ae05d2: Status 404 returned error can't find the container with id b370c88452268c3d7cb54dcf12872afd1ae627c89e61fa519ecd42fc93ae05d2 Apr 22 13:39:07.587455 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:07.587417 2579 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 13:39:08.589057 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:08.589017 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b79c7bb88-p8w9c" event={"ID":"615924fd-d8e9-48d2-8c72-6a4cb484cc10","Type":"ContainerStarted","Data":"030cbec5583f40c4f84b142497a998d8a4603bb9c5801672b29510b4d2b1fe65"} Apr 22 13:39:08.589057 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:08.589056 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b79c7bb88-p8w9c" event={"ID":"615924fd-d8e9-48d2-8c72-6a4cb484cc10","Type":"ContainerStarted","Data":"b370c88452268c3d7cb54dcf12872afd1ae627c89e61fa519ecd42fc93ae05d2"} Apr 22 13:39:08.618633 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:08.618589 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6b79c7bb88-p8w9c" podStartSLOduration=1.618575755 podStartE2EDuration="1.618575755s" podCreationTimestamp="2026-04-22 13:39:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 13:39:08.616669715 +0000 UTC m=+1066.553056617" watchObservedRunningTime="2026-04-22 13:39:08.618575755 +0000 UTC m=+1066.554962618" Apr 22 13:39:17.465600 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:17.465523 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:17.465600 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:17.465561 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:17.470193 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:17.470172 2579 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:39:17.618104 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:39:17.618082 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6b79c7bb88-p8w9c" Apr 22 13:41:22.548514 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:41:22.548407 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:41:22.554384 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:41:22.554365 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:46:22.569758 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:46:22.569646 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:46:22.573777 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:46:22.573529 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:51:22.588220 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:51:22.588113 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:51:22.592530 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:51:22.592512 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:56:22.608336 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:56:22.608240 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 13:56:22.612379 ip-10-0-128-54 kubenswrapper[2579]: I0422 13:56:22.612362 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 14:01:22.627556 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:01:22.627446 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 14:01:22.632484 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:01:22.632156 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 14:06:22.645600 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:06:22.645499 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 14:06:22.651295 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:06:22.651280 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 14:09:23.164922 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:23.164854 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-c7j5x_6560c958-0928-4fa1-8691-8fe812c8d951/global-pull-secret-syncer/0.log" Apr 22 14:09:23.264940 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:23.264915 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-hrtnk_87ad5783-96fb-40e6-abe4-0fda9e9532b6/konnectivity-agent/0.log" Apr 22 14:09:23.364198 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:23.364171 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-128-54.ec2.internal_16c68cd8f43f9f069938f5a899c05cc7/haproxy/0.log" Apr 22 14:09:26.245567 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.245537 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_443b8094-21ec-4417-9c5d-defe0b0a1326/alertmanager/0.log" Apr 22 14:09:26.275117 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.275093 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_443b8094-21ec-4417-9c5d-defe0b0a1326/config-reloader/0.log" Apr 22 14:09:26.317902 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.317880 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_443b8094-21ec-4417-9c5d-defe0b0a1326/kube-rbac-proxy-web/0.log" Apr 22 14:09:26.343436 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.343416 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_443b8094-21ec-4417-9c5d-defe0b0a1326/kube-rbac-proxy/0.log" Apr 22 14:09:26.374010 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.373992 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_443b8094-21ec-4417-9c5d-defe0b0a1326/kube-rbac-proxy-metric/0.log" Apr 22 14:09:26.403393 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.403371 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_443b8094-21ec-4417-9c5d-defe0b0a1326/prom-label-proxy/0.log" Apr 22 14:09:26.431923 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.431905 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_443b8094-21ec-4417-9c5d-defe0b0a1326/init-config-reloader/0.log" Apr 22 14:09:26.582310 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.582284 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_metrics-server-757fc4c546-cgzdq_e6e3682a-be46-4080-8f7f-01f99989110c/metrics-server/0.log" Apr 22 14:09:26.630722 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.630687 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-d5jgb_33daaf5e-222e-4197-a370-52af9fb7f55c/node-exporter/0.log" Apr 22 14:09:26.648233 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.648214 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-d5jgb_33daaf5e-222e-4197-a370-52af9fb7f55c/kube-rbac-proxy/0.log" Apr 22 14:09:26.666330 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.666315 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-d5jgb_33daaf5e-222e-4197-a370-52af9fb7f55c/init-textfile/0.log" Apr 22 14:09:26.841765 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.841710 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-rz469_9d530630-78b4-49c7-9015-fc89b9f15391/kube-rbac-proxy-main/0.log" Apr 22 14:09:26.863228 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.863203 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-rz469_9d530630-78b4-49c7-9015-fc89b9f15391/kube-rbac-proxy-self/0.log" Apr 22 14:09:26.882879 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.882863 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-9d44df66c-rz469_9d530630-78b4-49c7-9015-fc89b9f15391/openshift-state-metrics/0.log" Apr 22 14:09:26.912665 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.912648 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_148bd9c7-9015-47c6-9ee3-1fcb791544a6/prometheus/0.log" Apr 22 14:09:26.931350 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.931331 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_148bd9c7-9015-47c6-9ee3-1fcb791544a6/config-reloader/0.log" Apr 22 14:09:26.954666 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.954650 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_148bd9c7-9015-47c6-9ee3-1fcb791544a6/thanos-sidecar/0.log" Apr 22 14:09:26.974906 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.974888 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_148bd9c7-9015-47c6-9ee3-1fcb791544a6/kube-rbac-proxy-web/0.log" Apr 22 14:09:26.998063 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:26.998049 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_148bd9c7-9015-47c6-9ee3-1fcb791544a6/kube-rbac-proxy/0.log" Apr 22 14:09:27.017994 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:27.017979 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_148bd9c7-9015-47c6-9ee3-1fcb791544a6/kube-rbac-proxy-thanos/0.log" Apr 22 14:09:27.041127 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:27.041112 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_148bd9c7-9015-47c6-9ee3-1fcb791544a6/init-config-reloader/0.log" Apr 22 14:09:27.141726 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:27.141653 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-6cd4f57488-jvpmt_00f81ad6-66db-48b9-9348-edc079cfcd69/telemeter-client/0.log" Apr 22 14:09:27.164065 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:27.164044 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-6cd4f57488-jvpmt_00f81ad6-66db-48b9-9348-edc079cfcd69/reload/0.log" Apr 22 14:09:27.187788 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:27.187767 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-6cd4f57488-jvpmt_00f81ad6-66db-48b9-9348-edc079cfcd69/kube-rbac-proxy/0.log" Apr 22 14:09:29.299164 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:29.299142 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6b79c7bb88-p8w9c_615924fd-d8e9-48d2-8c72-6a4cb484cc10/console/0.log" Apr 22 14:09:30.337885 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.337847 2579 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g"] Apr 22 14:09:30.341318 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.341295 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.343597 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.343578 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-98vmz\"/\"openshift-service-ca.crt\"" Apr 22 14:09:30.343682 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.343579 2579 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-98vmz\"/\"kube-root-ca.crt\"" Apr 22 14:09:30.344700 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.344628 2579 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-98vmz\"/\"default-dockercfg-254cb\"" Apr 22 14:09:30.349772 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.349751 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g"] Apr 22 14:09:30.391364 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.391346 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-q8tfx_82906f03-65d1-460f-b410-9f8526350920/dns/0.log" Apr 22 14:09:30.409503 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.409475 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-q8tfx_82906f03-65d1-460f-b410-9f8526350920/kube-rbac-proxy/0.log" Apr 22 14:09:30.432232 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.432213 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-podres\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.432331 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.432262 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-proc\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.432331 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.432309 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-lib-modules\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.432404 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.432380 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-sys\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.432457 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.432401 2579 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ztvr\" (UniqueName: \"kubernetes.io/projected/4d0049af-9575-49a1-87d0-373f41ffea86-kube-api-access-2ztvr\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.476609 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.476583 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-5vr8r_c55e4ce3-f705-4675-8b13-55cf25d10b37/dns-node-resolver/0.log" Apr 22 14:09:30.533187 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533160 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-podres\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533295 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533211 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-proc\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533378 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533319 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-lib-modules\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533378 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533362 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-proc\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533378 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533325 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-podres\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533508 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533410 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-sys\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533508 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533457 2579 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2ztvr\" (UniqueName: \"kubernetes.io/projected/4d0049af-9575-49a1-87d0-373f41ffea86-kube-api-access-2ztvr\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533508 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533471 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-lib-modules\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.533508 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.533483 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4d0049af-9575-49a1-87d0-373f41ffea86-sys\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.540207 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.540190 2579 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ztvr\" (UniqueName: \"kubernetes.io/projected/4d0049af-9575-49a1-87d0-373f41ffea86-kube-api-access-2ztvr\") pod \"perf-node-gather-daemonset-hrl4g\" (UID: \"4d0049af-9575-49a1-87d0-373f41ffea86\") " pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.650589 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.650506 2579 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:30.760363 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.760192 2579 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g"] Apr 22 14:09:30.762490 ip-10-0-128-54 kubenswrapper[2579]: W0422 14:09:30.762463 2579 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4d0049af_9575_49a1_87d0_373f41ffea86.slice/crio-24b6062659510be33d4d7b3d08f58717d2269030333721f1208eb556658ff4ce WatchSource:0}: Error finding container 24b6062659510be33d4d7b3d08f58717d2269030333721f1208eb556658ff4ce: Status 404 returned error can't find the container with id 24b6062659510be33d4d7b3d08f58717d2269030333721f1208eb556658ff4ce Apr 22 14:09:30.764099 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.764083 2579 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 14:09:30.913148 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.913073 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_image-registry-756c9666d4-q6jfg_d02cc9bc-fe6e-44ad-a130-545bbcd2fa0f/registry/0.log" Apr 22 14:09:30.981106 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:30.981083 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-r48qq_2b26247c-8561-4d36-a271-45bd1c30c698/node-ca/0.log" Apr 22 14:09:31.458805 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:31.458773 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" event={"ID":"4d0049af-9575-49a1-87d0-373f41ffea86","Type":"ContainerStarted","Data":"a3aa19747b52bc1a7c11a99e3904d5da349e81b675eb8d60d5b80752b52ef582"} Apr 22 14:09:31.458805 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:31.458808 2579 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" event={"ID":"4d0049af-9575-49a1-87d0-373f41ffea86","Type":"ContainerStarted","Data":"24b6062659510be33d4d7b3d08f58717d2269030333721f1208eb556658ff4ce"} Apr 22 14:09:31.459260 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:31.458944 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:31.474467 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:31.474409 2579 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" podStartSLOduration=1.474391698 podStartE2EDuration="1.474391698s" podCreationTimestamp="2026-04-22 14:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 14:09:31.47307923 +0000 UTC m=+2889.409466103" watchObservedRunningTime="2026-04-22 14:09:31.474391698 +0000 UTC m=+2889.410778560" Apr 22 14:09:31.957047 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:31.957020 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-xg89z_52d135c0-aa2e-4533-986a-5002ff68781d/serve-healthcheck-canary/0.log" Apr 22 14:09:32.416039 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:32.416005 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-smwrp_512808ab-fbb1-46eb-86e9-a0fb60d532db/kube-rbac-proxy/0.log" Apr 22 14:09:32.433726 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:32.433700 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-smwrp_512808ab-fbb1-46eb-86e9-a0fb60d532db/exporter/0.log" Apr 22 14:09:32.452011 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:32.451994 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-smwrp_512808ab-fbb1-46eb-86e9-a0fb60d532db/extractor/0.log" Apr 22 14:09:34.001694 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:34.001669 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-jobset-operator_jobset-controller-manager-7876c9c48-fcxtp_f4bca08f-b935-4e6c-8f5b-3e16f5e93352/manager/0.log" Apr 22 14:09:37.471933 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:37.471897 2579 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-98vmz/perf-node-gather-daemonset-hrl4g" Apr 22 14:09:38.262401 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.262374 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-vzwxc_1d77049f-f54b-436c-8e41-5e0b8d54e8bf/kube-multus-additional-cni-plugins/0.log" Apr 22 14:09:38.281456 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.281413 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-vzwxc_1d77049f-f54b-436c-8e41-5e0b8d54e8bf/egress-router-binary-copy/0.log" Apr 22 14:09:38.301368 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.301350 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-vzwxc_1d77049f-f54b-436c-8e41-5e0b8d54e8bf/cni-plugins/0.log" Apr 22 14:09:38.319637 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.319614 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-vzwxc_1d77049f-f54b-436c-8e41-5e0b8d54e8bf/bond-cni-plugin/0.log" Apr 22 14:09:38.337943 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.337927 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-vzwxc_1d77049f-f54b-436c-8e41-5e0b8d54e8bf/routeoverride-cni/0.log" Apr 22 14:09:38.357561 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.357541 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-vzwxc_1d77049f-f54b-436c-8e41-5e0b8d54e8bf/whereabouts-cni-bincopy/0.log" Apr 22 14:09:38.376139 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.376121 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-vzwxc_1d77049f-f54b-436c-8e41-5e0b8d54e8bf/whereabouts-cni/0.log" Apr 22 14:09:38.424924 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.424886 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-grkk2_a4457d5d-3460-41bb-a335-8f9d5e2a4197/kube-multus/0.log" Apr 22 14:09:38.474535 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.474504 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-8vcg7_5dec8b65-6274-44b1-a169-ad20045acd1a/network-metrics-daemon/0.log" Apr 22 14:09:38.492597 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:38.492574 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-8vcg7_5dec8b65-6274-44b1-a169-ad20045acd1a/kube-rbac-proxy/0.log" Apr 22 14:09:39.716970 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.716864 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-controller/0.log" Apr 22 14:09:39.734716 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.734687 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/0.log" Apr 22 14:09:39.746112 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.746091 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovn-acl-logging/1.log" Apr 22 14:09:39.762613 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.762584 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/kube-rbac-proxy-node/0.log" Apr 22 14:09:39.788392 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.788362 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/kube-rbac-proxy-ovn-metrics/0.log" Apr 22 14:09:39.814132 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.814106 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/northd/0.log" Apr 22 14:09:39.837241 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.837216 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/nbdb/0.log" Apr 22 14:09:39.859884 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.859862 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/sbdb/0.log" Apr 22 14:09:39.951963 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:39.951933 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lqksf_782146a0-22aa-427e-aeb7-43bb6980c291/ovnkube-controller/0.log" Apr 22 14:09:41.321408 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:41.321332 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-tmp8r_4829e038-c6a2-4934-8182-fae81fb059a8/network-check-target-container/0.log" Apr 22 14:09:42.161371 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:42.161323 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-hv7gn_f757ef9d-b853-489b-9d90-79295660b525/iptables-alerter/0.log" Apr 22 14:09:42.763328 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:42.763302 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-wq59k_ee95ba9f-567a-461d-88c1-d3a904c366ad/tuned/0.log" Apr 22 14:09:46.325465 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:46.325419 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-qz6ss_ea7df7e2-bf24-4c31-92ce-352960562242/csi-driver/0.log" Apr 22 14:09:46.348067 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:46.348049 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-qz6ss_ea7df7e2-bf24-4c31-92ce-352960562242/csi-node-driver-registrar/0.log" Apr 22 14:09:46.372114 ip-10-0-128-54 kubenswrapper[2579]: I0422 14:09:46.372092 2579 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-qz6ss_ea7df7e2-bf24-4c31-92ce-352960562242/csi-liveness-probe/0.log"