Apr 17 09:20:50.070093 ip-10-0-129-70 systemd[1]: Starting Kubernetes Kubelet... Apr 17 09:20:50.511725 ip-10-0-129-70 kubenswrapper[2581]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 09:20:50.511725 ip-10-0-129-70 kubenswrapper[2581]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 17 09:20:50.511725 ip-10-0-129-70 kubenswrapper[2581]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 09:20:50.511725 ip-10-0-129-70 kubenswrapper[2581]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 17 09:20:50.511725 ip-10-0-129-70 kubenswrapper[2581]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 09:20:50.513046 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.512587 2581 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 17 09:20:50.516760 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516738 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:20:50.516760 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516755 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:20:50.516760 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516760 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:20:50.516760 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516764 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:20:50.516760 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516781 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516786 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516791 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516794 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516798 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516802 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516805 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516808 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516812 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516815 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516819 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516823 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516826 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516830 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516834 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516837 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516841 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516844 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516848 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516853 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:20:50.517065 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516856 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516872 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516879 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516883 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516888 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516893 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516897 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516901 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516905 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516909 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516914 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516919 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516924 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516929 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516934 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516939 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516945 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516951 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516956 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:20:50.517872 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516962 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516967 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516971 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516977 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516981 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516986 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516990 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516994 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.516998 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517003 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517007 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517011 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517015 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517019 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517023 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517027 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517032 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517036 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517041 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517045 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:20:50.518502 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517049 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517053 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517057 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517061 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517065 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517069 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517073 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517077 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517082 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517088 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517092 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517097 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517102 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517106 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517110 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517114 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517118 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517123 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517127 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517131 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:20:50.519003 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517135 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517139 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517143 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517752 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517761 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517766 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517790 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517795 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517800 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517804 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517809 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517814 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517818 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517822 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517826 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517831 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517835 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517839 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517843 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517848 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:20:50.519821 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517852 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517856 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517860 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517873 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517878 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517881 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517886 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517889 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517893 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517897 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517903 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517907 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517912 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517916 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517920 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517924 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517928 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517932 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517936 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517940 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:20:50.520636 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517944 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517948 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517952 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517956 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517960 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517964 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517968 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517973 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517977 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517981 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517985 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517989 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517993 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.517998 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518002 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518006 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518010 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518015 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518019 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:20:50.521141 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518023 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518029 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518033 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518038 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518042 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518046 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518051 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518055 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518059 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518063 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518068 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518072 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518077 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518082 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518087 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518091 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518095 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518098 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518102 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:20:50.521682 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518107 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518114 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518118 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518122 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518126 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518130 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518135 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518139 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518143 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518147 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.518151 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519105 2581 flags.go:64] FLAG: --address="0.0.0.0" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519126 2581 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519142 2581 flags.go:64] FLAG: --anonymous-auth="true" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519150 2581 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519156 2581 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519161 2581 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519168 2581 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519175 2581 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519180 2581 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519185 2581 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519190 2581 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 17 09:20:50.522242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519196 2581 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519201 2581 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519206 2581 flags.go:64] FLAG: --cgroup-root="" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519210 2581 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519216 2581 flags.go:64] FLAG: --client-ca-file="" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519220 2581 flags.go:64] FLAG: --cloud-config="" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519225 2581 flags.go:64] FLAG: --cloud-provider="external" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519229 2581 flags.go:64] FLAG: --cluster-dns="[]" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519244 2581 flags.go:64] FLAG: --cluster-domain="" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519248 2581 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519253 2581 flags.go:64] FLAG: --config-dir="" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519261 2581 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519267 2581 flags.go:64] FLAG: --container-log-max-files="5" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519291 2581 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519296 2581 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519302 2581 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519307 2581 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519312 2581 flags.go:64] FLAG: --contention-profiling="false" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519317 2581 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519322 2581 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519327 2581 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519332 2581 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519340 2581 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519345 2581 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519349 2581 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 17 09:20:50.522943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519354 2581 flags.go:64] FLAG: --enable-load-reader="false" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519359 2581 flags.go:64] FLAG: --enable-server="true" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519364 2581 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519372 2581 flags.go:64] FLAG: --event-burst="100" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519377 2581 flags.go:64] FLAG: --event-qps="50" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519382 2581 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519387 2581 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519393 2581 flags.go:64] FLAG: --eviction-hard="" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519400 2581 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519405 2581 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519410 2581 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519415 2581 flags.go:64] FLAG: --eviction-soft="" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519420 2581 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519425 2581 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519429 2581 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519434 2581 flags.go:64] FLAG: --experimental-mounter-path="" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519439 2581 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519444 2581 flags.go:64] FLAG: --fail-swap-on="true" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519450 2581 flags.go:64] FLAG: --feature-gates="" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519456 2581 flags.go:64] FLAG: --file-check-frequency="20s" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519461 2581 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519466 2581 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519471 2581 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519476 2581 flags.go:64] FLAG: --healthz-port="10248" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519482 2581 flags.go:64] FLAG: --help="false" Apr 17 09:20:50.523608 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519487 2581 flags.go:64] FLAG: --hostname-override="ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519492 2581 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519497 2581 flags.go:64] FLAG: --http-check-frequency="20s" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519502 2581 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519507 2581 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519514 2581 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519519 2581 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519524 2581 flags.go:64] FLAG: --image-service-endpoint="" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519528 2581 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519533 2581 flags.go:64] FLAG: --kube-api-burst="100" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519538 2581 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519543 2581 flags.go:64] FLAG: --kube-api-qps="50" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519547 2581 flags.go:64] FLAG: --kube-reserved="" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519553 2581 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519557 2581 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519563 2581 flags.go:64] FLAG: --kubelet-cgroups="" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519567 2581 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519572 2581 flags.go:64] FLAG: --lock-file="" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519577 2581 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519582 2581 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519587 2581 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519595 2581 flags.go:64] FLAG: --log-json-split-stream="false" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519600 2581 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519605 2581 flags.go:64] FLAG: --log-text-split-stream="false" Apr 17 09:20:50.524322 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519610 2581 flags.go:64] FLAG: --logging-format="text" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519617 2581 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519622 2581 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519627 2581 flags.go:64] FLAG: --manifest-url="" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519631 2581 flags.go:64] FLAG: --manifest-url-header="" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519638 2581 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519643 2581 flags.go:64] FLAG: --max-open-files="1000000" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519650 2581 flags.go:64] FLAG: --max-pods="110" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519655 2581 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519660 2581 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519665 2581 flags.go:64] FLAG: --memory-manager-policy="None" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519670 2581 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519674 2581 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519679 2581 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519686 2581 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519700 2581 flags.go:64] FLAG: --node-status-max-images="50" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519705 2581 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519709 2581 flags.go:64] FLAG: --oom-score-adj="-999" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519714 2581 flags.go:64] FLAG: --pod-cidr="" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519719 2581 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519729 2581 flags.go:64] FLAG: --pod-manifest-path="" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519735 2581 flags.go:64] FLAG: --pod-max-pids="-1" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519740 2581 flags.go:64] FLAG: --pods-per-core="0" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519745 2581 flags.go:64] FLAG: --port="10250" Apr 17 09:20:50.524938 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519750 2581 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519755 2581 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-02ebce6ea058318de" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519760 2581 flags.go:64] FLAG: --qos-reserved="" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519764 2581 flags.go:64] FLAG: --read-only-port="10255" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519785 2581 flags.go:64] FLAG: --register-node="true" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519791 2581 flags.go:64] FLAG: --register-schedulable="true" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519796 2581 flags.go:64] FLAG: --register-with-taints="" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519802 2581 flags.go:64] FLAG: --registry-burst="10" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519807 2581 flags.go:64] FLAG: --registry-qps="5" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519812 2581 flags.go:64] FLAG: --reserved-cpus="" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519820 2581 flags.go:64] FLAG: --reserved-memory="" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519826 2581 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519831 2581 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519835 2581 flags.go:64] FLAG: --rotate-certificates="false" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519840 2581 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519844 2581 flags.go:64] FLAG: --runonce="false" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519849 2581 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519854 2581 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519859 2581 flags.go:64] FLAG: --seccomp-default="false" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519864 2581 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519869 2581 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519874 2581 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519879 2581 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519885 2581 flags.go:64] FLAG: --storage-driver-password="root" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519890 2581 flags.go:64] FLAG: --storage-driver-secure="false" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519895 2581 flags.go:64] FLAG: --storage-driver-table="stats" Apr 17 09:20:50.525543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519901 2581 flags.go:64] FLAG: --storage-driver-user="root" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519906 2581 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519910 2581 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519915 2581 flags.go:64] FLAG: --system-cgroups="" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519920 2581 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519931 2581 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519936 2581 flags.go:64] FLAG: --tls-cert-file="" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519941 2581 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519948 2581 flags.go:64] FLAG: --tls-min-version="" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519952 2581 flags.go:64] FLAG: --tls-private-key-file="" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519957 2581 flags.go:64] FLAG: --topology-manager-policy="none" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519962 2581 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519967 2581 flags.go:64] FLAG: --topology-manager-scope="container" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519971 2581 flags.go:64] FLAG: --v="2" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519979 2581 flags.go:64] FLAG: --version="false" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519985 2581 flags.go:64] FLAG: --vmodule="" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519994 2581 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.519999 2581 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520149 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520156 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520160 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520164 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520168 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520172 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:20:50.526165 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520177 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520181 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520185 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520190 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520194 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520198 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520203 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520207 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520211 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520215 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520219 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520225 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520229 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520233 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520237 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520241 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520245 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520249 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520254 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520258 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:20:50.526790 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520262 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520266 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520270 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520274 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520281 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520285 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520289 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520293 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520299 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520306 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520312 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520318 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520323 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520328 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520332 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520337 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520342 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520348 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520353 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:20:50.527323 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520358 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520363 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520368 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520372 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520378 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520383 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520387 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520392 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520396 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520400 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520405 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520409 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520413 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520417 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520421 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520425 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520430 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520436 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520440 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520444 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:20:50.527818 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520448 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520452 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520456 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520460 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520464 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520479 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520484 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520488 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520491 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520496 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520500 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520505 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520509 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520513 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520517 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520521 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520527 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520531 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520535 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520539 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:20:50.528320 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.520544 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.521300 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.528635 2581 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.528652 2581 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528698 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528702 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528705 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528708 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528711 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528714 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528717 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528720 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528723 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528725 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528728 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528730 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:20:50.528830 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528733 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528735 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528738 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528741 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528743 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528746 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528749 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528751 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528754 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528757 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528760 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528763 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528765 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528780 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528783 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528786 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528791 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528794 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528798 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528801 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:20:50.529248 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528804 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528807 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528810 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528813 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528815 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528818 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528821 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528823 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528826 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528828 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528831 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528834 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528836 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528839 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528842 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528845 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528848 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528851 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528853 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:20:50.529733 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528856 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528858 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528861 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528863 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528866 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528868 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528871 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528874 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528876 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528879 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528882 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528884 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528888 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528891 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528893 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528896 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528898 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528900 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528903 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528905 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:20:50.530199 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528907 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528910 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528913 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528915 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528917 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528920 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528922 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528925 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528927 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528930 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528932 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528935 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528937 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528941 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.528951 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:20:50.530669 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.528957 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529052 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529056 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529059 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529062 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529064 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529067 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529069 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529072 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529074 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529077 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529080 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529083 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529085 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529088 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529090 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529092 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529095 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529097 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529100 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 09:20:50.531044 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529102 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529105 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529107 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529110 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529112 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529115 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529117 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529120 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529122 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529125 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529127 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529129 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529132 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529134 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529137 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529139 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529143 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529147 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529150 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529152 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 09:20:50.531562 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529156 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529158 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529161 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529164 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529167 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529170 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529173 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529175 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529179 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529182 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529184 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529187 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529189 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529192 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529194 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529197 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529200 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529202 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529205 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529207 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 09:20:50.532070 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529209 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529212 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529214 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529217 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529220 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529222 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529225 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529227 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529229 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529233 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529236 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529239 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529242 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529244 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529247 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529249 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529252 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529255 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529258 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 09:20:50.532598 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529260 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529263 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529265 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529268 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529270 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529272 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529275 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:50.529277 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.529282 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 09:20:50.533062 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.530304 2581 server.go:962] "Client rotation is on, will bootstrap in background" Apr 17 09:20:50.533277 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.533149 2581 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 17 09:20:50.534061 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.534049 2581 server.go:1019] "Starting client certificate rotation" Apr 17 09:20:50.534158 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.534143 2581 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 17 09:20:50.534249 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.534184 2581 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 17 09:20:50.557606 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.557587 2581 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 17 09:20:50.561541 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.561516 2581 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 17 09:20:50.576800 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.576782 2581 log.go:25] "Validated CRI v1 runtime API" Apr 17 09:20:50.582305 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.582290 2581 log.go:25] "Validated CRI v1 image API" Apr 17 09:20:50.584003 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.583981 2581 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 17 09:20:50.588748 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.588725 2581 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 17 09:20:50.590333 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.590311 2581 fs.go:135] Filesystem UUIDs: map[1895bf5d-b9ca-4562-a1b0-3d1c91274803:/dev/nvme0n1p3 415dae42-43d2-46e4-b9de-1a993b64ef97:/dev/nvme0n1p4 7B77-95E7:/dev/nvme0n1p2] Apr 17 09:20:50.590393 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.590330 2581 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 17 09:20:50.596570 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.596441 2581 manager.go:217] Machine: {Timestamp:2026-04-17 09:20:50.594403407 +0000 UTC m=+0.407316572 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3230059 MemoryCapacity:33164488704 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2250752e5a8fbe902b6b10bda1aeff SystemUUID:ec225075-2e5a-8fbe-902b-6b10bda1aeff BootID:3918db15-6fc4-4939-8fa7-0cdc4961224b Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16582242304 Type:vfs Inodes:4048399 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6632898560 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16582246400 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:74:dc:76:b8:c1 Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:74:dc:76:b8:c1 Speed:0 Mtu:9001} {Name:ovs-system MacAddress:7a:a0:a0:1e:d1:df Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33164488704 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 17 09:20:50.596570 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.596563 2581 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 17 09:20:50.596693 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.596644 2581 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 17 09:20:50.597800 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.597763 2581 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 17 09:20:50.597929 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.597801 2581 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-129-70.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 17 09:20:50.597977 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.597940 2581 topology_manager.go:138] "Creating topology manager with none policy" Apr 17 09:20:50.597977 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.597949 2581 container_manager_linux.go:306] "Creating device plugin manager" Apr 17 09:20:50.597977 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.597962 2581 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 17 09:20:50.598647 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.598638 2581 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 17 09:20:50.599521 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.599510 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 17 09:20:50.599642 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.599633 2581 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 17 09:20:50.602562 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.602525 2581 kubelet.go:491] "Attempting to sync node with API server" Apr 17 09:20:50.602612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.602565 2581 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 17 09:20:50.602612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.602585 2581 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 17 09:20:50.602612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.602594 2581 kubelet.go:397] "Adding apiserver pod source" Apr 17 09:20:50.602745 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.602661 2581 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 17 09:20:50.603824 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.603813 2581 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 17 09:20:50.603864 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.603832 2581 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 17 09:20:50.607800 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.607724 2581 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 17 09:20:50.609623 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.609601 2581 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 17 09:20:50.611328 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611309 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 17 09:20:50.611398 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611335 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 17 09:20:50.611398 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611345 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 17 09:20:50.611398 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611354 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 17 09:20:50.611398 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611364 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 17 09:20:50.611398 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611377 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 17 09:20:50.611398 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611392 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 17 09:20:50.611398 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611398 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 17 09:20:50.611591 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611405 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 17 09:20:50.611591 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611412 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 17 09:20:50.611591 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611423 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 17 09:20:50.611591 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.611433 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 17 09:20:50.612349 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.612336 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 17 09:20:50.612380 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.612354 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 17 09:20:50.615975 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.615960 2581 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 17 09:20:50.616042 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.616006 2581 server.go:1295] "Started kubelet" Apr 17 09:20:50.616161 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.616112 2581 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 17 09:20:50.616247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.616120 2581 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 17 09:20:50.616247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.616201 2581 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 17 09:20:50.616827 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.616800 2581 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-129-70.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 17 09:20:50.616904 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.616883 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-129-70.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 17 09:20:50.616947 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.616889 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 17 09:20:50.617049 ip-10-0-129-70 systemd[1]: Started Kubernetes Kubelet. Apr 17 09:20:50.617287 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.617267 2581 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 17 09:20:50.618144 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.618129 2581 server.go:317] "Adding debug handlers to kubelet server" Apr 17 09:20:50.622549 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.622532 2581 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 17 09:20:50.622647 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.622556 2581 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 17 09:20:50.623289 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.623248 2581 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 17 09:20:50.623289 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.623251 2581 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 17 09:20:50.623435 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.623298 2581 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 17 09:20:50.623435 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.623385 2581 reconstruct.go:97] "Volume reconstruction finished" Apr 17 09:20:50.623435 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.623396 2581 reconciler.go:26] "Reconciler: start to sync state" Apr 17 09:20:50.623435 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.623423 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:50.625310 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.625279 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 17 09:20:50.626309 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.625351 2581 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-129-70.ec2.internal.18a71a70fdb2f97a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-129-70.ec2.internal,UID:ip-10-0-129-70.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-10-0-129-70.ec2.internal,},FirstTimestamp:2026-04-17 09:20:50.615974266 +0000 UTC m=+0.428887433,LastTimestamp:2026-04-17 09:20:50.615974266 +0000 UTC m=+0.428887433,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-129-70.ec2.internal,}" Apr 17 09:20:50.627152 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627138 2581 factory.go:153] Registering CRI-O factory Apr 17 09:20:50.627248 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627163 2581 factory.go:223] Registration of the crio container factory successfully Apr 17 09:20:50.627248 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627218 2581 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 17 09:20:50.627248 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627227 2581 factory.go:55] Registering systemd factory Apr 17 09:20:50.627248 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627244 2581 factory.go:223] Registration of the systemd container factory successfully Apr 17 09:20:50.627434 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627289 2581 factory.go:103] Registering Raw factory Apr 17 09:20:50.627434 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627303 2581 manager.go:1196] Started watching for new ooms in manager Apr 17 09:20:50.627932 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.627919 2581 manager.go:319] Starting recovery of all containers Apr 17 09:20:50.630561 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.630535 2581 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"ip-10-0-129-70.ec2.internal\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Apr 17 09:20:50.632936 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.632900 2581 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-25vmm" Apr 17 09:20:50.639686 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.639558 2581 manager.go:324] Recovery completed Apr 17 09:20:50.640084 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.640067 2581 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-25vmm" Apr 17 09:20:50.644492 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.644466 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:20:50.646897 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.646882 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:20:50.646950 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.646910 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:20:50.646950 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.646922 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:20:50.647408 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.647396 2581 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 17 09:20:50.647408 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.647406 2581 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 17 09:20:50.647493 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.647420 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 17 09:20:50.648938 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.648861 2581 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-129-70.ec2.internal.18a71a70ff8accc2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-129-70.ec2.internal,UID:ip-10-0-129-70.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node ip-10-0-129-70.ec2.internal status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:ip-10-0-129-70.ec2.internal,},FirstTimestamp:2026-04-17 09:20:50.64689581 +0000 UTC m=+0.459808975,LastTimestamp:2026-04-17 09:20:50.64689581 +0000 UTC m=+0.459808975,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-129-70.ec2.internal,}" Apr 17 09:20:50.649582 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.649570 2581 policy_none.go:49] "None policy: Start" Apr 17 09:20:50.649630 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.649585 2581 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 17 09:20:50.649630 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.649595 2581 state_mem.go:35] "Initializing new in-memory state store" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.685049 2581 manager.go:341] "Starting Device Plugin manager" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.685084 2581 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.685097 2581 server.go:85] "Starting device plugin registration server" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.685333 2581 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.685345 2581 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.685445 2581 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.685524 2581 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.685531 2581 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.686026 2581 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 17 09:20:50.687160 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.686061 2581 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:50.746683 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.746647 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 17 09:20:50.747900 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.747882 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 17 09:20:50.747972 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.747918 2581 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 17 09:20:50.747972 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.747941 2581 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 17 09:20:50.747972 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.747951 2581 kubelet.go:2451] "Starting kubelet main sync loop" Apr 17 09:20:50.748107 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.747992 2581 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 17 09:20:50.749911 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.749889 2581 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:20:50.785507 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.785438 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:20:50.786389 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.786373 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:20:50.786467 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.786401 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:20:50.786467 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.786411 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:20:50.786467 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.786432 2581 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.792357 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.792338 2581 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.792447 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.792363 2581 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-129-70.ec2.internal\": node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:50.817853 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.817827 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:50.848562 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.848539 2581 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal"] Apr 17 09:20:50.848666 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.848602 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:20:50.849500 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.849487 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:20:50.849551 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.849510 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:20:50.849551 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.849522 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:20:50.850743 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.850732 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:20:50.850898 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.850884 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.850939 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.850912 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:20:50.851487 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.851464 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:20:50.851487 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.851474 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:20:50.851487 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.851491 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:20:50.851678 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.851495 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:20:50.851678 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.851505 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:20:50.851678 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.851506 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:20:50.852611 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.852596 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.852659 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.852624 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 09:20:50.853302 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.853282 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientMemory" Apr 17 09:20:50.853302 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.853302 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 09:20:50.853423 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.853323 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeHasSufficientPID" Apr 17 09:20:50.875342 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.875322 2581 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-129-70.ec2.internal\" not found" node="ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.879179 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.879165 2581 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-129-70.ec2.internal\" not found" node="ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.918529 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:50.918510 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:50.925479 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.925466 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/6413b1b3363a0f5c8e5c6748344609d0-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal\" (UID: \"6413b1b3363a0f5c8e5c6748344609d0\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.925535 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.925488 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6413b1b3363a0f5c8e5c6748344609d0-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal\" (UID: \"6413b1b3363a0f5c8e5c6748344609d0\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:50.925535 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:50.925505 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/9af7c40fd2295f75411f633f4223f981-config\") pod \"kube-apiserver-proxy-ip-10-0-129-70.ec2.internal\" (UID: \"9af7c40fd2295f75411f633f4223f981\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.018808 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.018764 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.026196 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.026178 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/6413b1b3363a0f5c8e5c6748344609d0-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal\" (UID: \"6413b1b3363a0f5c8e5c6748344609d0\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.026298 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.026207 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6413b1b3363a0f5c8e5c6748344609d0-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal\" (UID: \"6413b1b3363a0f5c8e5c6748344609d0\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.026298 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.026224 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/9af7c40fd2295f75411f633f4223f981-config\") pod \"kube-apiserver-proxy-ip-10-0-129-70.ec2.internal\" (UID: \"9af7c40fd2295f75411f633f4223f981\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.026298 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.026262 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6413b1b3363a0f5c8e5c6748344609d0-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal\" (UID: \"6413b1b3363a0f5c8e5c6748344609d0\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.026298 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.026271 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/6413b1b3363a0f5c8e5c6748344609d0-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal\" (UID: \"6413b1b3363a0f5c8e5c6748344609d0\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.026412 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.026308 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/9af7c40fd2295f75411f633f4223f981-config\") pod \"kube-apiserver-proxy-ip-10-0-129-70.ec2.internal\" (UID: \"9af7c40fd2295f75411f633f4223f981\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.119656 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.119589 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.177011 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.176973 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.181673 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.181656 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" Apr 17 09:20:51.220287 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.220262 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.320936 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.320902 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.421534 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.421461 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.452091 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.452066 2581 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:20:51.522595 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.522559 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.533954 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.533939 2581 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 17 09:20:51.534084 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.534066 2581 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 17 09:20:51.534121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.534111 2581 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 17 09:20:51.623016 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.622987 2581 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 17 09:20:51.623016 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.622987 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.629998 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.629977 2581 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:20:51.639344 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.639320 2581 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 17 09:20:51.642129 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.642106 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-16 09:15:50 +0000 UTC" deadline="2027-11-17 02:00:17.879564017 +0000 UTC" Apr 17 09:20:51.642129 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.642128 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="13888h39m26.237438307s" Apr 17 09:20:51.661571 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.661552 2581 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-4zhq8" Apr 17 09:20:51.669694 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.669680 2581 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-4zhq8" Apr 17 09:20:51.723386 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.723360 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.824211 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.824177 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:51.863917 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:51.863873 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6413b1b3363a0f5c8e5c6748344609d0.slice/crio-430aa16ccbf7011e94bac98a3e8399dc479a6d81b6158fca78d96b5612f90456 WatchSource:0}: Error finding container 430aa16ccbf7011e94bac98a3e8399dc479a6d81b6158fca78d96b5612f90456: Status 404 returned error can't find the container with id 430aa16ccbf7011e94bac98a3e8399dc479a6d81b6158fca78d96b5612f90456 Apr 17 09:20:51.870286 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:51.870272 2581 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 09:20:51.888683 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:51.888661 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9af7c40fd2295f75411f633f4223f981.slice/crio-a0c7e43153dca9552c19c42343f81605f3c3db59b30e71ed165c51919562f4fc WatchSource:0}: Error finding container a0c7e43153dca9552c19c42343f81605f3c3db59b30e71ed165c51919562f4fc: Status 404 returned error can't find the container with id a0c7e43153dca9552c19c42343f81605f3c3db59b30e71ed165c51919562f4fc Apr 17 09:20:51.924767 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:51.924745 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:52.025255 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.025231 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:52.125985 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.125954 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-129-70.ec2.internal\" not found" Apr 17 09:20:52.189049 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.189022 2581 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:20:52.223153 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.223124 2581 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" Apr 17 09:20:52.236668 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.236622 2581 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 17 09:20:52.237574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.237561 2581 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" Apr 17 09:20:52.246219 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.246197 2581 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 17 09:20:52.603828 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.603743 2581 apiserver.go:52] "Watching apiserver" Apr 17 09:20:52.610435 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.610411 2581 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 17 09:20:52.610895 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.610865 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/konnectivity-agent-gdhfx","kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm","openshift-cluster-node-tuning-operator/tuned-9t4gl","openshift-dns/node-resolver-6h6cx","openshift-image-registry/node-ca-sj6zp","openshift-multus/multus-additional-cni-plugins-2h6z9","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal","openshift-multus/multus-s8tqz","openshift-multus/network-metrics-daemon-84m5n","openshift-network-diagnostics/network-check-target-zmtbt","openshift-network-operator/iptables-alerter-8r786","openshift-ovn-kubernetes/ovnkube-node-64h7n","kube-system/global-pull-secret-syncer-hbnq9"] Apr 17 09:20:52.615564 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.615538 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:52.615695 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.615641 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:20:52.615765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.615699 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.617874 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.617840 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.618598 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.618522 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 17 09:20:52.618598 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.618568 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 17 09:20:52.618782 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.618527 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 17 09:20:52.618782 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.618727 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-fc98p\"" Apr 17 09:20:52.618782 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.618743 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 17 09:20:52.620143 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.620067 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-m5qnc\"" Apr 17 09:20:52.620143 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.620075 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 17 09:20:52.620697 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.620415 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 17 09:20:52.620697 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.620436 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 17 09:20:52.622359 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.622331 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.624260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.624233 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 17 09:20:52.624672 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.624653 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 17 09:20:52.624672 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.624668 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.624833 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.624700 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-4xddh\"" Apr 17 09:20:52.626639 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.626613 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-5zcxm\"" Apr 17 09:20:52.626639 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.626635 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 17 09:20:52.626998 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.626983 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 17 09:20:52.627696 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.627639 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.630057 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.629974 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 17 09:20:52.630057 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.630030 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-ldmbl\"" Apr 17 09:20:52.630336 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.630298 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 17 09:20:52.631125 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.630044 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.631125 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.630838 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 17 09:20:52.631397 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.631132 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.633050 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.632900 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 17 09:20:52.635090 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.633537 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-s2p4j\"" Apr 17 09:20:52.635090 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.634034 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 17 09:20:52.635090 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.634042 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 17 09:20:52.635090 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.634844 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 17 09:20:52.635351 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.635221 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 17 09:20:52.635651 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.635633 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 17 09:20:52.636242 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636223 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:20:52.636636 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636368 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-gw65g\"" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636723 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-etc-selinux\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636847 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-cnibin\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636886 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636935 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-os-release\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636969 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.636975 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-hostroot\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637004 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6c744de0-b054-441d-a935-6063dac1784d-multus-daemon-config\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637035 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9qsx\" (UniqueName: \"kubernetes.io/projected/6c744de0-b054-441d-a935-6063dac1784d-kube-api-access-x9qsx\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637066 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-registration-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637105 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-sys-fs\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637138 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k88qz\" (UniqueName: \"kubernetes.io/projected/bd655abe-79ad-487e-adbe-b874b344be93-kube-api-access-k88qz\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637169 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637182 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-host\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637224 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c744de0-b054-441d-a935-6063dac1784d-cni-binary-copy\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637253 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-netns\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637277 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-etc-kubernetes\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637305 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.637716 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637356 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/56322e26-7d25-4937-af43-be08d98d2866-hosts-file\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.638421 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637387 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4rgn\" (UniqueName: \"kubernetes.io/projected/56322e26-7d25-4937-af43-be08d98d2866-kube-api-access-l4rgn\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.638421 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637439 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-serviceca\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.638421 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637465 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-cni-multus\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.638421 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637529 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-device-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.638421 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637599 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs6cd\" (UniqueName: \"kubernetes.io/projected/729c1d11-856e-4ad5-8a1c-46175de73a63-kube-api-access-cs6cd\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.638421 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.637631 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-os-release\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.638731 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.638711 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-cni-binary-copy\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.638807 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.638749 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/56322e26-7d25-4937-af43-be08d98d2866-tmp-dir\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.638807 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.638802 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-socket-dir-parent\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.638910 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.638827 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-k8s-cni-cncf-io\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.638910 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.638849 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-cni-bin\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.639004 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.638929 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-socket-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.639057 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.638969 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.639057 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639039 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f98v\" (UniqueName: \"kubernetes.io/projected/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-kube-api-access-7f98v\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.639152 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639065 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x842f\" (UniqueName: \"kubernetes.io/projected/504207a8-f3c9-45e3-a2cd-66458cf6eff8-kube-api-access-x842f\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:52.639152 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639107 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-cni-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.639152 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639130 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-cnibin\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.639317 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639221 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-kubelet\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.639317 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639288 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-conf-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.639317 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639312 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-multus-certs\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.639470 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639351 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-system-cni-dir\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.639470 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639374 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-kubelet-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.639470 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639453 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 17 09:20:52.639605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639491 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-tvg9s\"" Apr 17 09:20:52.639657 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639630 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.639705 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639660 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-system-cni-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.639816 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.639799 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 17 09:20:52.641392 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.641375 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:52.641470 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.641443 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:20:52.643733 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.643707 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.646104 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.645940 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.646104 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.646028 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:20:52.646252 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.646152 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 17 09:20:52.646328 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.646314 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-nfd8z\"" Apr 17 09:20:52.646385 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.646372 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 17 09:20:52.646385 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.646377 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 17 09:20:52.671675 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.671498 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-16 09:15:51 +0000 UTC" deadline="2027-11-30 13:58:14.807662053 +0000 UTC" Apr 17 09:20:52.671675 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.671522 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14212h37m22.136143762s" Apr 17 09:20:52.724463 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.724442 2581 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 17 09:20:52.740425 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740401 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-systemd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.740555 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740431 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysconfig\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.740555 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740465 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:52.740555 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740543 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysctl-d\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.740694 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740581 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-systemd\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.740694 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740595 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-lib-modules\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.740694 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740614 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:52.740694 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740662 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-registration-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.740889 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740730 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-registration-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.740889 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740743 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-ovn\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.740889 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740798 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-var-lib-kubelet\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.740889 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.740825 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:52.740889 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740829 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-k88qz\" (UniqueName: \"kubernetes.io/projected/bd655abe-79ad-487e-adbe-b874b344be93-kube-api-access-k88qz\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.740889 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740855 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-host\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740904 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-host\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.740918 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:20:53.240868881 +0000 UTC m=+3.053782037 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740943 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-netns\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.740981 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72r4z\" (UniqueName: \"kubernetes.io/projected/6160dcfd-b407-4815-bb22-a36a6d52799d-kube-api-access-72r4z\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741007 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-log-socket\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741032 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741062 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/56322e26-7d25-4937-af43-be08d98d2866-hosts-file\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741061 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-netns\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741088 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-serviceca\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741112 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-cni-multus\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741115 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/56322e26-7d25-4937-af43-be08d98d2866-hosts-file\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.741164 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741154 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-device-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741164 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-cni-multus\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741197 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-run-ovn-kubernetes\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741234 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-env-overrides\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741224 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-device-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741265 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-os-release\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741289 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-cni-binary-copy\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741315 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-socket-dir-parent\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741339 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-socket-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741361 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-os-release\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741370 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/31f8d7d4-b363-48ef-b10f-1ad81e1b496d-agent-certs\") pod \"konnectivity-agent-gdhfx\" (UID: \"31f8d7d4-b363-48ef-b10f-1ad81e1b496d\") " pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741415 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741459 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7f98v\" (UniqueName: \"kubernetes.io/projected/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-kube-api-access-7f98v\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741496 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-socket-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.741695 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741512 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-serviceca\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741554 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x842f\" (UniqueName: \"kubernetes.io/projected/504207a8-f3c9-45e3-a2cd-66458cf6eff8-kube-api-access-x842f\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.741994 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-conf-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742032 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-multus-certs\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742067 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/819cb802-41c1-426d-bcc4-c96157d9e5ed-dbus\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742098 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742129 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-slash\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742149 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-cni-bin\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742169 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovnkube-config\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742199 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-run\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742227 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-host\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.742260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742248 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/31f8d7d4-b363-48ef-b10f-1ad81e1b496d-konnectivity-ca\") pod \"konnectivity-agent-gdhfx\" (UID: \"31f8d7d4-b363-48ef-b10f-1ad81e1b496d\") " pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.742721 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742281 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.742721 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742314 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-cni-netd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.742721 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742355 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovn-node-metrics-cert\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.742721 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742392 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-system-cni-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.742721 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742429 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-etc-selinux\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.742721 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742457 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.743040 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742754 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-conf-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.743040 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.742819 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-multus-certs\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.743130 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743049 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-socket-dir-parent\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.743130 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743077 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-system-cni-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.743130 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743079 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.743390 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743351 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-node-log\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.743390 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743362 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.743548 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743366 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-etc-selinux\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743431 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-cnibin\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744160 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-os-release\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744195 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-hostroot\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744227 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6c744de0-b054-441d-a935-6063dac1784d-multus-daemon-config\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744257 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x9qsx\" (UniqueName: \"kubernetes.io/projected/6c744de0-b054-441d-a935-6063dac1784d-kube-api-access-x9qsx\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744293 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-sys-fs\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744327 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-622dh\" (UniqueName: \"kubernetes.io/projected/fd8d642b-833b-4048-a36c-25491ebf64f9-kube-api-access-622dh\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744358 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-var-lib-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744389 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c744de0-b054-441d-a935-6063dac1784d-cni-binary-copy\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744433 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-etc-kubernetes\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744459 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-kubelet\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744505 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-etc-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744535 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-kubernetes\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744562 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysctl-conf\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744594 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744622 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l4rgn\" (UniqueName: \"kubernetes.io/projected/56322e26-7d25-4937-af43-be08d98d2866-kube-api-access-l4rgn\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744655 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cs6cd\" (UniqueName: \"kubernetes.io/projected/729c1d11-856e-4ad5-8a1c-46175de73a63-kube-api-access-cs6cd\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744685 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/56322e26-7d25-4937-af43-be08d98d2866-tmp-dir\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744714 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-k8s-cni-cncf-io\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744746 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-cni-bin\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744789 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-systemd-units\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744826 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-run-netns\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743474 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-cnibin\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.743633 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd655abe-79ad-487e-adbe-b874b344be93-cni-binary-copy\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744896 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovnkube-script-lib\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744913 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-run-k8s-cni-cncf-io\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.744959 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-cni-bin\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745079 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-kube-api-access-qv6fd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745106 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-sys\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745130 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-cni-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745163 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-cnibin\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745351 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/56322e26-7d25-4937-af43-be08d98d2866-tmp-dir\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745461 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6c744de0-b054-441d-a935-6063dac1784d-multus-daemon-config\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.745574 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745537 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745596 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-sys-fs\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745615 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-os-release\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745674 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-hostroot\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745682 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-kubelet\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745697 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-multus-cni-dir\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745714 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-etc-kubernetes\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745741 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-host-var-lib-kubelet\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745753 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c744de0-b054-441d-a935-6063dac1784d-cnibin\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745760 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/6160dcfd-b407-4815-bb22-a36a6d52799d-iptables-alerter-script\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745829 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-modprobe-d\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745867 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-tuned\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745896 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fd8d642b-833b-4048-a36c-25491ebf64f9-tmp\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.745989 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-system-cni-dir\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.746011 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd655abe-79ad-487e-adbe-b874b344be93-system-cni-dir\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.746092 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-kubelet-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.746103 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c744de0-b054-441d-a935-6063dac1784d-cni-binary-copy\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.746254 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.746165 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6160dcfd-b407-4815-bb22-a36a6d52799d-host-slash\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.746890 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.746196 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c1d11-856e-4ad5-8a1c-46175de73a63-kubelet-dir\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.746890 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.746230 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/819cb802-41c1-426d-bcc4-c96157d9e5ed-kubelet-config\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.751952 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.751917 2581 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 17 09:20:52.752378 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.752312 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" event={"ID":"9af7c40fd2295f75411f633f4223f981","Type":"ContainerStarted","Data":"a0c7e43153dca9552c19c42343f81605f3c3db59b30e71ed165c51919562f4fc"} Apr 17 09:20:52.753800 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.753758 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" event={"ID":"6413b1b3363a0f5c8e5c6748344609d0","Type":"ContainerStarted","Data":"430aa16ccbf7011e94bac98a3e8399dc479a6d81b6158fca78d96b5612f90456"} Apr 17 09:20:52.755758 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.755740 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f98v\" (UniqueName: \"kubernetes.io/projected/9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31-kube-api-access-7f98v\") pod \"node-ca-sj6zp\" (UID: \"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31\") " pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.755852 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.755763 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9qsx\" (UniqueName: \"kubernetes.io/projected/6c744de0-b054-441d-a935-6063dac1784d-kube-api-access-x9qsx\") pod \"multus-s8tqz\" (UID: \"6c744de0-b054-441d-a935-6063dac1784d\") " pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.755852 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.755790 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-k88qz\" (UniqueName: \"kubernetes.io/projected/bd655abe-79ad-487e-adbe-b874b344be93-kube-api-access-k88qz\") pod \"multus-additional-cni-plugins-2h6z9\" (UID: \"bd655abe-79ad-487e-adbe-b874b344be93\") " pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.755852 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.755790 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4rgn\" (UniqueName: \"kubernetes.io/projected/56322e26-7d25-4937-af43-be08d98d2866-kube-api-access-l4rgn\") pod \"node-resolver-6h6cx\" (UID: \"56322e26-7d25-4937-af43-be08d98d2866\") " pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.756207 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.756186 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs6cd\" (UniqueName: \"kubernetes.io/projected/729c1d11-856e-4ad5-8a1c-46175de73a63-kube-api-access-cs6cd\") pod \"aws-ebs-csi-driver-node-jqwtm\" (UID: \"729c1d11-856e-4ad5-8a1c-46175de73a63\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.756873 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.756849 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x842f\" (UniqueName: \"kubernetes.io/projected/504207a8-f3c9-45e3-a2cd-66458cf6eff8-kube-api-access-x842f\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:52.847223 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847193 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-slash\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847223 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847228 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-cni-bin\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847304 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-cni-bin\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847307 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-slash\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847348 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovnkube-config\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847367 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-run\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847382 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-host\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847398 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/31f8d7d4-b363-48ef-b10f-1ad81e1b496d-konnectivity-ca\") pod \"konnectivity-agent-gdhfx\" (UID: \"31f8d7d4-b363-48ef-b10f-1ad81e1b496d\") " pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847435 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-run\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847443 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-cni-netd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847473 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847468 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-host\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847513 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-cni-netd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847473 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovn-node-metrics-cert\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847557 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847581 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-node-log\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847609 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-622dh\" (UniqueName: \"kubernetes.io/projected/fd8d642b-833b-4048-a36c-25491ebf64f9-kube-api-access-622dh\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847641 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-var-lib-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847667 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-kubelet\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847691 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-etc-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.847704 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847714 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-kubernetes\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847736 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysctl-conf\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.847786 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret podName:819cb802-41c1-426d-bcc4-c96157d9e5ed nodeName:}" failed. No retries permitted until 2026-04-17 09:20:53.347753839 +0000 UTC m=+3.160667014 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret") pod "global-pull-secret-syncer-hbnq9" (UID: "819cb802-41c1-426d-bcc4-c96157d9e5ed") : object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847804 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-systemd-units\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847828 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-run-netns\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847849 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovnkube-script-lib\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.847905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847897 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-systemd-units\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847892 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysctl-conf\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847939 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-kube-api-access-qv6fd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847958 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-node-log\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847968 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-sys\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.847998 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/6160dcfd-b407-4815-bb22-a36a6d52799d-iptables-alerter-script\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848025 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-modprobe-d\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848037 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovnkube-config\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848050 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-tuned\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848053 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/31f8d7d4-b363-48ef-b10f-1ad81e1b496d-konnectivity-ca\") pod \"konnectivity-agent-gdhfx\" (UID: \"31f8d7d4-b363-48ef-b10f-1ad81e1b496d\") " pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848074 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fd8d642b-833b-4048-a36c-25491ebf64f9-tmp\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848111 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6160dcfd-b407-4815-bb22-a36a6d52799d-host-slash\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848159 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/819cb802-41c1-426d-bcc4-c96157d9e5ed-kubelet-config\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848197 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-systemd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848222 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysconfig\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848256 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848285 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysctl-d\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.848605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848286 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-var-lib-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848326 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-systemd\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848341 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovnkube-script-lib\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848353 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-lib-modules\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848377 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-run-netns\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848398 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/819cb802-41c1-426d-bcc4-c96157d9e5ed-kubelet-config\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848391 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-ovn\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848432 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysctl-d\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848436 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-ovn\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848435 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-var-lib-kubelet\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848488 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-var-lib-kubelet\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848501 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-sysconfig\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848491 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-72r4z\" (UniqueName: \"kubernetes.io/projected/6160dcfd-b407-4815-bb22-a36a6d52799d-kube-api-access-72r4z\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848534 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-systemd\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848534 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-log-socket\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848539 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-sys\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848562 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-log-socket\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848577 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.849348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848602 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-systemd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848433 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-kubernetes\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848612 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-run-ovn-kubernetes\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848637 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-env-overrides\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848649 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-lib-modules\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848651 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-etc-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848490 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6160dcfd-b407-4815-bb22-a36a6d52799d-host-slash\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848675 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/31f8d7d4-b363-48ef-b10f-1ad81e1b496d-agent-certs\") pod \"konnectivity-agent-gdhfx\" (UID: \"31f8d7d4-b363-48ef-b10f-1ad81e1b496d\") " pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848721 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/819cb802-41c1-426d-bcc4-c96157d9e5ed-dbus\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848743 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-modprobe-d\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848752 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848799 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848844 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-run-openvswitch\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848677 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-kubelet\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.848905 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-host-run-ovn-kubernetes\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.849027 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/819cb802-41c1-426d-bcc4-c96157d9e5ed-dbus\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.849130 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-env-overrides\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.849298 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/6160dcfd-b407-4815-bb22-a36a6d52799d-iptables-alerter-script\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.850765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.850682 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-ovn-node-metrics-cert\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.850920 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.850901 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/fd8d642b-833b-4048-a36c-25491ebf64f9-etc-tuned\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.850969 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.850945 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/fd8d642b-833b-4048-a36c-25491ebf64f9-tmp\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.851267 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.851248 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/31f8d7d4-b363-48ef-b10f-1ad81e1b496d-agent-certs\") pod \"konnectivity-agent-gdhfx\" (UID: \"31f8d7d4-b363-48ef-b10f-1ad81e1b496d\") " pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.856185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.856117 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:20:52.856185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.856143 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:20:52.856185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.856155 2581 projected.go:194] Error preparing data for projected volume kube-api-access-md7x7 for pod openshift-network-diagnostics/network-check-target-zmtbt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:52.856375 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:52.856232 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7 podName:02351e99-a2a3-489f-b4dc-f8f12e779f0d nodeName:}" failed. No retries permitted until 2026-04-17 09:20:53.356215732 +0000 UTC m=+3.169128909 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-md7x7" (UniqueName: "kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7") pod "network-check-target-zmtbt" (UID: "02351e99-a2a3-489f-b4dc-f8f12e779f0d") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:52.858605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.858572 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/604b7c24-c2c4-45ab-9dbe-a714dd18cd14-kube-api-access-qv6fd\") pod \"ovnkube-node-64h7n\" (UID: \"604b7c24-c2c4-45ab-9dbe-a714dd18cd14\") " pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.859262 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.859228 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-622dh\" (UniqueName: \"kubernetes.io/projected/fd8d642b-833b-4048-a36c-25491ebf64f9-kube-api-access-622dh\") pod \"tuned-9t4gl\" (UID: \"fd8d642b-833b-4048-a36c-25491ebf64f9\") " pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.859364 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.859228 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-72r4z\" (UniqueName: \"kubernetes.io/projected/6160dcfd-b407-4815-bb22-a36a6d52799d-kube-api-access-72r4z\") pod \"iptables-alerter-8r786\" (UID: \"6160dcfd-b407-4815-bb22-a36a6d52799d\") " pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:52.927584 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.927546 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-s8tqz" Apr 17 09:20:52.936403 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.936370 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" Apr 17 09:20:52.947247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.947216 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" Apr 17 09:20:52.955011 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.954986 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6h6cx" Apr 17 09:20:52.960593 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.960572 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sj6zp" Apr 17 09:20:52.968222 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.968202 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:20:52.975851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.975834 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" Apr 17 09:20:52.984402 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.984381 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:20:52.988243 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.988227 2581 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:20:52.992325 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:52.992309 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-8r786" Apr 17 09:20:53.252080 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.252042 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:53.252236 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.252195 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:53.252272 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.252254 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:20:54.252239161 +0000 UTC m=+4.065152313 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:53.353215 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.353171 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:53.353408 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.353335 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:53.353408 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.353397 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret podName:819cb802-41c1-426d-bcc4-c96157d9e5ed nodeName:}" failed. No retries permitted until 2026-04-17 09:20:54.353382612 +0000 UTC m=+4.166295764 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret") pod "global-pull-secret-syncer-hbnq9" (UID: "819cb802-41c1-426d-bcc4-c96157d9e5ed") : object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:53.453962 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.453931 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:53.454102 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.454079 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:20:53.454102 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.454099 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:20:53.454165 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.454109 2581 projected.go:194] Error preparing data for projected volume kube-api-access-md7x7 for pod openshift-network-diagnostics/network-check-target-zmtbt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:53.454165 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:53.454156 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7 podName:02351e99-a2a3-489f-b4dc-f8f12e779f0d nodeName:}" failed. No retries permitted until 2026-04-17 09:20:54.454142346 +0000 UTC m=+4.267055498 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-md7x7" (UniqueName: "kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7") pod "network-check-target-zmtbt" (UID: "02351e99-a2a3-489f-b4dc-f8f12e779f0d") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:53.490838 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.490808 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b8cdaba_1bb6_47fd_b3c8_300ebe7fff31.slice/crio-f4dc7080c9ebe2a23cb848a3868dede0cf7c394421436507ab3fe9fc383ac237 WatchSource:0}: Error finding container f4dc7080c9ebe2a23cb848a3868dede0cf7c394421436507ab3fe9fc383ac237: Status 404 returned error can't find the container with id f4dc7080c9ebe2a23cb848a3868dede0cf7c394421436507ab3fe9fc383ac237 Apr 17 09:20:53.492149 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.492131 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd655abe_79ad_487e_adbe_b874b344be93.slice/crio-d3efb13bf75499c3374d89ae600d97264c4d7e5210264b49fe5037e3750e0964 WatchSource:0}: Error finding container d3efb13bf75499c3374d89ae600d97264c4d7e5210264b49fe5037e3750e0964: Status 404 returned error can't find the container with id d3efb13bf75499c3374d89ae600d97264c4d7e5210264b49fe5037e3750e0964 Apr 17 09:20:53.494971 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.494947 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod729c1d11_856e_4ad5_8a1c_46175de73a63.slice/crio-a3e863238d0cc36c8dcc7c80636d2e5dc485b587777b0933f2a15d9d1047ce55 WatchSource:0}: Error finding container a3e863238d0cc36c8dcc7c80636d2e5dc485b587777b0933f2a15d9d1047ce55: Status 404 returned error can't find the container with id a3e863238d0cc36c8dcc7c80636d2e5dc485b587777b0933f2a15d9d1047ce55 Apr 17 09:20:53.496012 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.495967 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod604b7c24_c2c4_45ab_9dbe_a714dd18cd14.slice/crio-e1da532960c3cee4441110a1ed35323cc5b72312a3cc9169b57c53716d89b8ba WatchSource:0}: Error finding container e1da532960c3cee4441110a1ed35323cc5b72312a3cc9169b57c53716d89b8ba: Status 404 returned error can't find the container with id e1da532960c3cee4441110a1ed35323cc5b72312a3cc9169b57c53716d89b8ba Apr 17 09:20:53.496917 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.496894 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c744de0_b054_441d_a935_6063dac1784d.slice/crio-c4733f843376cdb0ce1e9509a3f81b76527d79f97fcf05c266b42ae6ac166f6d WatchSource:0}: Error finding container c4733f843376cdb0ce1e9509a3f81b76527d79f97fcf05c266b42ae6ac166f6d: Status 404 returned error can't find the container with id c4733f843376cdb0ce1e9509a3f81b76527d79f97fcf05c266b42ae6ac166f6d Apr 17 09:20:53.497991 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.497958 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd8d642b_833b_4048_a36c_25491ebf64f9.slice/crio-2f62c5ffc1fd55228417673d039255dd0248c390dbbfe358c22072b0fe7c7eee WatchSource:0}: Error finding container 2f62c5ffc1fd55228417673d039255dd0248c390dbbfe358c22072b0fe7c7eee: Status 404 returned error can't find the container with id 2f62c5ffc1fd55228417673d039255dd0248c390dbbfe358c22072b0fe7c7eee Apr 17 09:20:53.499930 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.499912 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6160dcfd_b407_4815_bb22_a36a6d52799d.slice/crio-3136f2aea59a76fb4d5b176694fa886f0172d39067aa4a3549e80cfe4df04414 WatchSource:0}: Error finding container 3136f2aea59a76fb4d5b176694fa886f0172d39067aa4a3549e80cfe4df04414: Status 404 returned error can't find the container with id 3136f2aea59a76fb4d5b176694fa886f0172d39067aa4a3549e80cfe4df04414 Apr 17 09:20:53.500701 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:20:53.500539 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31f8d7d4_b363_48ef_b10f_1ad81e1b496d.slice/crio-f6c0fa21887fe1380b25f9780faf2887c1ae72dac67f2ffc4ebdbe4e26126719 WatchSource:0}: Error finding container f6c0fa21887fe1380b25f9780faf2887c1ae72dac67f2ffc4ebdbe4e26126719: Status 404 returned error can't find the container with id f6c0fa21887fe1380b25f9780faf2887c1ae72dac67f2ffc4ebdbe4e26126719 Apr 17 09:20:53.671851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.671665 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-16 09:15:51 +0000 UTC" deadline="2028-01-20 11:00:31.833917032 +0000 UTC" Apr 17 09:20:53.671851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.671842 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="15433h39m38.162078227s" Apr 17 09:20:53.755749 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.755650 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-gdhfx" event={"ID":"31f8d7d4-b363-48ef-b10f-1ad81e1b496d","Type":"ContainerStarted","Data":"f6c0fa21887fe1380b25f9780faf2887c1ae72dac67f2ffc4ebdbe4e26126719"} Apr 17 09:20:53.756558 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.756529 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-8r786" event={"ID":"6160dcfd-b407-4815-bb22-a36a6d52799d","Type":"ContainerStarted","Data":"3136f2aea59a76fb4d5b176694fa886f0172d39067aa4a3549e80cfe4df04414"} Apr 17 09:20:53.757649 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.757630 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" event={"ID":"fd8d642b-833b-4048-a36c-25491ebf64f9","Type":"ContainerStarted","Data":"2f62c5ffc1fd55228417673d039255dd0248c390dbbfe358c22072b0fe7c7eee"} Apr 17 09:20:53.758521 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.758505 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" event={"ID":"729c1d11-856e-4ad5-8a1c-46175de73a63","Type":"ContainerStarted","Data":"a3e863238d0cc36c8dcc7c80636d2e5dc485b587777b0933f2a15d9d1047ce55"} Apr 17 09:20:53.759497 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.759476 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sj6zp" event={"ID":"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31","Type":"ContainerStarted","Data":"f4dc7080c9ebe2a23cb848a3868dede0cf7c394421436507ab3fe9fc383ac237"} Apr 17 09:20:53.760407 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.760383 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s8tqz" event={"ID":"6c744de0-b054-441d-a935-6063dac1784d","Type":"ContainerStarted","Data":"c4733f843376cdb0ce1e9509a3f81b76527d79f97fcf05c266b42ae6ac166f6d"} Apr 17 09:20:53.761381 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.761352 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"e1da532960c3cee4441110a1ed35323cc5b72312a3cc9169b57c53716d89b8ba"} Apr 17 09:20:53.762362 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.762331 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerStarted","Data":"d3efb13bf75499c3374d89ae600d97264c4d7e5210264b49fe5037e3750e0964"} Apr 17 09:20:53.763931 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.763913 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" event={"ID":"9af7c40fd2295f75411f633f4223f981","Type":"ContainerStarted","Data":"64ced520a69f945b42dbe1f700a009d46b014ee83cf9cd0710eb3aa0a0b35b68"} Apr 17 09:20:53.764886 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.764865 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6h6cx" event={"ID":"56322e26-7d25-4937-af43-be08d98d2866","Type":"ContainerStarted","Data":"4abd289a9e3946d3ef2ae223d656e44f96abcb2dd46d6d7848bd481369bf1576"} Apr 17 09:20:53.777527 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:53.777484 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-129-70.ec2.internal" podStartSLOduration=1.777474522 podStartE2EDuration="1.777474522s" podCreationTimestamp="2026-04-17 09:20:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:20:53.777466464 +0000 UTC m=+3.590379637" watchObservedRunningTime="2026-04-17 09:20:53.777474522 +0000 UTC m=+3.590387696" Apr 17 09:20:54.260543 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.259247 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:54.260543 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.259445 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:54.260543 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.259513 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:20:56.259493474 +0000 UTC m=+6.072406628 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:54.360458 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.360370 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:54.360625 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.360585 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:54.360686 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.360646 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret podName:819cb802-41c1-426d-bcc4-c96157d9e5ed nodeName:}" failed. No retries permitted until 2026-04-17 09:20:56.360627416 +0000 UTC m=+6.173540574 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret") pod "global-pull-secret-syncer-hbnq9" (UID: "819cb802-41c1-426d-bcc4-c96157d9e5ed") : object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:54.462280 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.461700 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:54.462280 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.461867 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:20:54.462280 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.461886 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:20:54.462280 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.461898 2581 projected.go:194] Error preparing data for projected volume kube-api-access-md7x7 for pod openshift-network-diagnostics/network-check-target-zmtbt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:54.462280 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.461953 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7 podName:02351e99-a2a3-489f-b4dc-f8f12e779f0d nodeName:}" failed. No retries permitted until 2026-04-17 09:20:56.461935624 +0000 UTC m=+6.274848782 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-md7x7" (UniqueName: "kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7") pod "network-check-target-zmtbt" (UID: "02351e99-a2a3-489f-b4dc-f8f12e779f0d") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:54.598188 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.598159 2581 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 09:20:54.751785 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.748353 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:54.751785 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.748491 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:20:54.751785 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.748989 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:54.751785 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.749080 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:20:54.751785 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.751630 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:54.751785 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:54.751719 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:20:54.800168 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.799418 2581 generic.go:358] "Generic (PLEG): container finished" podID="6413b1b3363a0f5c8e5c6748344609d0" containerID="b8d9cbcfbab9e8eb87b9d7da522663a51891e90c0717f314f7f7589a8b9fa949" exitCode=0 Apr 17 09:20:54.800168 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:54.799936 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" event={"ID":"6413b1b3363a0f5c8e5c6748344609d0","Type":"ContainerDied","Data":"b8d9cbcfbab9e8eb87b9d7da522663a51891e90c0717f314f7f7589a8b9fa949"} Apr 17 09:20:55.840498 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:55.840445 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" event={"ID":"6413b1b3363a0f5c8e5c6748344609d0","Type":"ContainerStarted","Data":"80075d96f2f4da1e1961f4aab1521482a22bc2b8cabb3c50d351d9db4222b3b0"} Apr 17 09:20:55.855996 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:55.855943 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-129-70.ec2.internal" podStartSLOduration=3.855924366 podStartE2EDuration="3.855924366s" podCreationTimestamp="2026-04-17 09:20:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:20:55.854895656 +0000 UTC m=+5.667808834" watchObservedRunningTime="2026-04-17 09:20:55.855924366 +0000 UTC m=+5.668837542" Apr 17 09:20:56.282030 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:56.281341 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:56.282030 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.281547 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:56.282030 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.281612 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:00.281594295 +0000 UTC m=+10.094507454 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:20:56.382319 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:56.382279 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:56.382496 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.382485 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:56.382555 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.382546 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret podName:819cb802-41c1-426d-bcc4-c96157d9e5ed nodeName:}" failed. No retries permitted until 2026-04-17 09:21:00.382529288 +0000 UTC m=+10.195442446 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret") pod "global-pull-secret-syncer-hbnq9" (UID: "819cb802-41c1-426d-bcc4-c96157d9e5ed") : object "kube-system"/"original-pull-secret" not registered Apr 17 09:20:56.483062 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:56.483020 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:56.483255 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.483213 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:20:56.483255 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.483243 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:20:56.483255 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.483256 2581 projected.go:194] Error preparing data for projected volume kube-api-access-md7x7 for pod openshift-network-diagnostics/network-check-target-zmtbt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:56.483442 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.483332 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7 podName:02351e99-a2a3-489f-b4dc-f8f12e779f0d nodeName:}" failed. No retries permitted until 2026-04-17 09:21:00.483313161 +0000 UTC m=+10.296226336 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-md7x7" (UniqueName: "kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7") pod "network-check-target-zmtbt" (UID: "02351e99-a2a3-489f-b4dc-f8f12e779f0d") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:20:56.750407 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:56.749948 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:56.750407 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:56.749945 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:56.750407 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.750068 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:20:56.750407 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.750174 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:20:56.750407 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:56.750221 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:56.750407 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:56.750285 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:20:58.748618 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:58.748341 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:20:58.748618 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:58.748357 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:20:58.748618 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:58.748474 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:20:58.748618 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:20:58.748547 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:20:58.749201 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:58.748675 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:20:58.749201 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:20:58.748791 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:00.315722 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:00.315656 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:00.316183 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.315824 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:00.316183 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.315900 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:08.315880622 +0000 UTC m=+18.128793779 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:00.416781 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:00.416737 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:00.416930 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.416881 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 17 09:21:00.416976 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.416956 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret podName:819cb802-41c1-426d-bcc4-c96157d9e5ed nodeName:}" failed. No retries permitted until 2026-04-17 09:21:08.416940279 +0000 UTC m=+18.229853437 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret") pod "global-pull-secret-syncer-hbnq9" (UID: "819cb802-41c1-426d-bcc4-c96157d9e5ed") : object "kube-system"/"original-pull-secret" not registered Apr 17 09:21:00.518179 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:00.518145 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:00.518355 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.518329 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:00.518355 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.518354 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:00.518461 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.518370 2581 projected.go:194] Error preparing data for projected volume kube-api-access-md7x7 for pod openshift-network-diagnostics/network-check-target-zmtbt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:00.518461 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.518453 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7 podName:02351e99-a2a3-489f-b4dc-f8f12e779f0d nodeName:}" failed. No retries permitted until 2026-04-17 09:21:08.518414382 +0000 UTC m=+18.331327541 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-md7x7" (UniqueName: "kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7") pod "network-check-target-zmtbt" (UID: "02351e99-a2a3-489f-b4dc-f8f12e779f0d") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:00.749432 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:00.749393 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:00.749601 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.749508 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:00.749677 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:00.749600 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:00.749760 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.749734 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:00.749846 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:00.749792 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:00.749902 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:00.749874 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:02.748842 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.748574 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:02.749222 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:02.748941 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:02.749222 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.748655 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:02.749222 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:02.749045 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:02.749222 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.748574 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:02.749222 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:02.749121 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:02.853688 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.853594 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" event={"ID":"729c1d11-856e-4ad5-8a1c-46175de73a63","Type":"ContainerStarted","Data":"8240d28488d814e05d50b566cbf2370a2a384dc0d756d6ac10de456c9dd54f6a"} Apr 17 09:21:02.855012 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.854986 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sj6zp" event={"ID":"9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31","Type":"ContainerStarted","Data":"217f5c92d3363c2c0edf89da0a089d06a3abee8c8ad3b974602c8d0ee8179f7f"} Apr 17 09:21:02.856532 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.856504 2581 generic.go:358] "Generic (PLEG): container finished" podID="bd655abe-79ad-487e-adbe-b874b344be93" containerID="d0d60593be5e17d444c42e49384507af1624a17451eedb352ca64cd4ffb31b3f" exitCode=0 Apr 17 09:21:02.856648 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.856591 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerDied","Data":"d0d60593be5e17d444c42e49384507af1624a17451eedb352ca64cd4ffb31b3f"} Apr 17 09:21:02.858027 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.857924 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6h6cx" event={"ID":"56322e26-7d25-4937-af43-be08d98d2866","Type":"ContainerStarted","Data":"8868aed460025e4a7a226857fadf9b588c5205613ac714b9c047266008e5b755"} Apr 17 09:21:02.859454 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.859412 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-gdhfx" event={"ID":"31f8d7d4-b363-48ef-b10f-1ad81e1b496d","Type":"ContainerStarted","Data":"027bec5d175aeb228a9d12a2060f791e3ebaa41d0152979e5db8a5be15b022e4"} Apr 17 09:21:02.860849 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.860829 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" event={"ID":"fd8d642b-833b-4048-a36c-25491ebf64f9","Type":"ContainerStarted","Data":"6f627df07650a10afdcabf0a9692cc9a13e0c16977610bd64ef21822c9ec7c11"} Apr 17 09:21:02.869674 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.869626 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-sj6zp" podStartSLOduration=4.344818774 podStartE2EDuration="12.869614116s" podCreationTimestamp="2026-04-17 09:20:50 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.493189194 +0000 UTC m=+3.306102356" lastFinishedPulling="2026-04-17 09:21:02.017984526 +0000 UTC m=+11.830897698" observedRunningTime="2026-04-17 09:21:02.869464558 +0000 UTC m=+12.682377723" watchObservedRunningTime="2026-04-17 09:21:02.869614116 +0000 UTC m=+12.682527290" Apr 17 09:21:02.885130 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.885075 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-9t4gl" podStartSLOduration=3.341056821 podStartE2EDuration="11.885061095s" podCreationTimestamp="2026-04-17 09:20:51 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.500065218 +0000 UTC m=+3.312978383" lastFinishedPulling="2026-04-17 09:21:02.04406949 +0000 UTC m=+11.856982657" observedRunningTime="2026-04-17 09:21:02.884544576 +0000 UTC m=+12.697457751" watchObservedRunningTime="2026-04-17 09:21:02.885061095 +0000 UTC m=+12.697974269" Apr 17 09:21:02.928529 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.928473 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-gdhfx" podStartSLOduration=3.405517932 podStartE2EDuration="11.928442868s" podCreationTimestamp="2026-04-17 09:20:51 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.504002823 +0000 UTC m=+3.316915979" lastFinishedPulling="2026-04-17 09:21:02.026927757 +0000 UTC m=+11.839840915" observedRunningTime="2026-04-17 09:21:02.898833861 +0000 UTC m=+12.711747035" watchObservedRunningTime="2026-04-17 09:21:02.928442868 +0000 UTC m=+12.741356045" Apr 17 09:21:02.944789 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:02.944717 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6h6cx" podStartSLOduration=4.420585197 podStartE2EDuration="12.944700097s" podCreationTimestamp="2026-04-17 09:20:50 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.503670922 +0000 UTC m=+3.316584080" lastFinishedPulling="2026-04-17 09:21:02.027785827 +0000 UTC m=+11.840698980" observedRunningTime="2026-04-17 09:21:02.944037667 +0000 UTC m=+12.756950843" watchObservedRunningTime="2026-04-17 09:21:02.944700097 +0000 UTC m=+12.757613284" Apr 17 09:21:03.112400 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:03.112325 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:21:03.112923 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:03.112902 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:21:03.864589 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:03.864553 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-8r786" event={"ID":"6160dcfd-b407-4815-bb22-a36a6d52799d","Type":"ContainerStarted","Data":"5fac86380ccd1a92966e35d87b88c3a47d2123b1f9dcb4682f2837f5d585a501"} Apr 17 09:21:03.879537 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:03.879488 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-8r786" podStartSLOduration=4.354930979 podStartE2EDuration="12.87947322s" podCreationTimestamp="2026-04-17 09:20:51 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.502379615 +0000 UTC m=+3.315292767" lastFinishedPulling="2026-04-17 09:21:02.026921849 +0000 UTC m=+11.839835008" observedRunningTime="2026-04-17 09:21:03.879167351 +0000 UTC m=+13.692080524" watchObservedRunningTime="2026-04-17 09:21:03.87947322 +0000 UTC m=+13.692386395" Apr 17 09:21:04.748585 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:04.748549 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:04.748786 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:04.748549 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:04.748786 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:04.748673 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:04.748786 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:04.748555 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:04.748928 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:04.748760 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:04.748928 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:04.748895 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:04.865752 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:04.865720 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 17 09:21:06.748319 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:06.748280 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:06.749025 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:06.748332 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:06.749025 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:06.748295 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:06.749025 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:06.748409 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:06.749025 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:06.748529 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:06.749025 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:06.748619 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:08.376450 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:08.376411 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:08.377150 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.376590 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:08.377150 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.376659 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:24.376640345 +0000 UTC m=+34.189553501 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:08.477079 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:08.477041 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:08.477240 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.477202 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 17 09:21:08.477307 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.477283 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret podName:819cb802-41c1-426d-bcc4-c96157d9e5ed nodeName:}" failed. No retries permitted until 2026-04-17 09:21:24.477263045 +0000 UTC m=+34.290176213 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret") pod "global-pull-secret-syncer-hbnq9" (UID: "819cb802-41c1-426d-bcc4-c96157d9e5ed") : object "kube-system"/"original-pull-secret" not registered Apr 17 09:21:08.578396 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:08.578356 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:08.578582 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.578553 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:08.578582 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.578581 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:08.578678 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.578594 2581 projected.go:194] Error preparing data for projected volume kube-api-access-md7x7 for pod openshift-network-diagnostics/network-check-target-zmtbt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:08.578678 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.578657 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7 podName:02351e99-a2a3-489f-b4dc-f8f12e779f0d nodeName:}" failed. No retries permitted until 2026-04-17 09:21:24.578638653 +0000 UTC m=+34.391551819 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-md7x7" (UniqueName: "kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7") pod "network-check-target-zmtbt" (UID: "02351e99-a2a3-489f-b4dc-f8f12e779f0d") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:08.749186 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:08.749148 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:08.749348 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.749273 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:08.749348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:08.749285 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:08.749348 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:08.749317 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:08.749507 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.749375 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:08.749507 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:08.749447 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:10.749017 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:10.748979 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:10.749556 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:10.749074 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:10.749556 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:10.749158 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:10.749556 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:10.749258 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:10.749556 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:10.749323 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:10.749556 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:10.749346 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:12.748748 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:12.748514 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:12.749340 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:12.748522 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:12.749340 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:12.748969 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:12.749340 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:12.748858 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:12.749340 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:12.748540 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:12.749340 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:12.749060 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:13.215823 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.215597 2581 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 17 09:21:13.699130 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.699011 2581 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-17T09:21:13.215818794Z","UUID":"add4764b-7f3d-4447-8d71-d4fc31350f93","Handler":null,"Name":"","Endpoint":""} Apr 17 09:21:13.700544 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.700527 2581 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 17 09:21:13.700544 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.700549 2581 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 17 09:21:13.885070 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.885036 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"29e25a5b0e550be5b40d4518cb88fd77cc6186ab385108d0ea98a9d2a436414c"} Apr 17 09:21:13.885070 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.885077 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"d77f215bf671af13bc68f66ec304c890184a269f75452faeb8182f8c923fcb98"} Apr 17 09:21:13.885711 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.885093 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"c2698759adff956cbe55a8e746013a1a569e6e83d8441a70100f9c01fcaaa224"} Apr 17 09:21:13.885711 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.885105 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"fa1d8fbd68623cd665b35b9897d7afe25c3a49e32ba089ea37e1f92be4d21972"} Apr 17 09:21:13.885711 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.885118 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"ea2cccc0c8acc3b6c6749fdd46b9628e60490fb1e5246519871a874eef98b7a1"} Apr 17 09:21:13.885711 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.885130 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"2e9e743d911003a40e88d19a5c7678dfa6abed7a8ec4398e813d75514803e23a"} Apr 17 09:21:13.886942 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.886919 2581 generic.go:358] "Generic (PLEG): container finished" podID="bd655abe-79ad-487e-adbe-b874b344be93" containerID="53f5912f17cc140a010decc87b6fe4e1c2625c9fe80b909cd14c842fe3111e01" exitCode=0 Apr 17 09:21:13.887096 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.887009 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerDied","Data":"53f5912f17cc140a010decc87b6fe4e1c2625c9fe80b909cd14c842fe3111e01"} Apr 17 09:21:13.888745 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.888722 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" event={"ID":"729c1d11-856e-4ad5-8a1c-46175de73a63","Type":"ContainerStarted","Data":"6dc8f96474fd8020fa4c61647448952d6447e9847ab2f385e9e209b37798a788"} Apr 17 09:21:13.890498 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.890385 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s8tqz" event={"ID":"6c744de0-b054-441d-a935-6063dac1784d","Type":"ContainerStarted","Data":"7402c5f56bb2cfc4e48450ac9d6d817e6dbeb227936b002091b74e937ba52886"} Apr 17 09:21:13.922332 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:13.922282 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-s8tqz" podStartSLOduration=4.36010321 podStartE2EDuration="23.922270532s" podCreationTimestamp="2026-04-17 09:20:50 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.498633291 +0000 UTC m=+3.311546448" lastFinishedPulling="2026-04-17 09:21:13.060800612 +0000 UTC m=+22.873713770" observedRunningTime="2026-04-17 09:21:13.921711263 +0000 UTC m=+23.734624437" watchObservedRunningTime="2026-04-17 09:21:13.922270532 +0000 UTC m=+23.735183705" Apr 17 09:21:14.748699 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:14.748675 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:14.748835 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:14.748676 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:14.748835 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:14.748800 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:14.748946 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:14.748676 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:14.748946 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:14.748874 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:14.749036 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:14.748968 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:14.894184 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:14.894100 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" event={"ID":"729c1d11-856e-4ad5-8a1c-46175de73a63","Type":"ContainerStarted","Data":"2fb0f818a813b7f67c9c60c3e0510c25f84e7b41e3d1856ad621e34496eb7c06"} Apr 17 09:21:14.910355 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:14.910311 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-jqwtm" podStartSLOduration=3.812728964 podStartE2EDuration="24.910296402s" podCreationTimestamp="2026-04-17 09:20:50 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.496873989 +0000 UTC m=+3.309787142" lastFinishedPulling="2026-04-17 09:21:14.594441359 +0000 UTC m=+24.407354580" observedRunningTime="2026-04-17 09:21:14.909840663 +0000 UTC m=+24.722753837" watchObservedRunningTime="2026-04-17 09:21:14.910296402 +0000 UTC m=+24.723209601" Apr 17 09:21:15.130279 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:15.130246 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:21:15.130440 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:15.130376 2581 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 17 09:21:15.130754 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:15.130740 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-gdhfx" Apr 17 09:21:15.901596 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:15.901503 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"3bc2779df5f2412b9bbc98cbf6c107d0536113673d7b39a3325e0574335b6380"} Apr 17 09:21:15.903244 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:15.903213 2581 generic.go:358] "Generic (PLEG): container finished" podID="bd655abe-79ad-487e-adbe-b874b344be93" containerID="9ca83eb4d99b433d62232300cc2c1927d45373b337e99216dbaccaaa96fed00a" exitCode=0 Apr 17 09:21:15.903365 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:15.903300 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerDied","Data":"9ca83eb4d99b433d62232300cc2c1927d45373b337e99216dbaccaaa96fed00a"} Apr 17 09:21:16.748356 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:16.748323 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:16.748356 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:16.748349 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:16.748605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:16.748330 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:16.748605 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:16.748456 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:16.748605 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:16.748573 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:16.748747 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:16.748628 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:17.910178 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.909988 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" event={"ID":"604b7c24-c2c4-45ab-9dbe-a714dd18cd14","Type":"ContainerStarted","Data":"aa3124eddbe63d2e975fe9b5122f702ad0207af65ab8e68c70ba83c5d8208600"} Apr 17 09:21:17.910831 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.910208 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:21:17.910831 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.910226 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:21:17.910831 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.910239 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:21:17.912514 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.912482 2581 generic.go:358] "Generic (PLEG): container finished" podID="bd655abe-79ad-487e-adbe-b874b344be93" containerID="6b883071fde28cae047b073dccfe65f0b2929c704da215a9f45f49b52bcefc9e" exitCode=0 Apr 17 09:21:17.912639 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.912537 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerDied","Data":"6b883071fde28cae047b073dccfe65f0b2929c704da215a9f45f49b52bcefc9e"} Apr 17 09:21:17.925157 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.925134 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:21:17.925304 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.925258 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:21:17.935609 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:17.935564 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" podStartSLOduration=7.37257756 podStartE2EDuration="26.935551927s" podCreationTimestamp="2026-04-17 09:20:51 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.497882404 +0000 UTC m=+3.310795557" lastFinishedPulling="2026-04-17 09:21:13.060856772 +0000 UTC m=+22.873769924" observedRunningTime="2026-04-17 09:21:17.935097482 +0000 UTC m=+27.748010667" watchObservedRunningTime="2026-04-17 09:21:17.935551927 +0000 UTC m=+27.748465103" Apr 17 09:21:18.748229 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:18.748187 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:18.748410 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:18.748187 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:18.748410 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:18.748357 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:18.748410 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:18.748203 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:18.748568 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:18.748419 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:18.748568 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:18.748515 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:19.700732 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:19.700701 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-hbnq9"] Apr 17 09:21:19.701120 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:19.700812 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:19.701120 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:19.700888 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:19.704349 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:19.704320 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-zmtbt"] Apr 17 09:21:19.704505 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:19.704418 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:19.704570 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:19.704528 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:19.704937 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:19.704910 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-84m5n"] Apr 17 09:21:19.705007 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:19.704995 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:19.705137 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:19.705104 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:20.752616 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:20.751304 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:20.752616 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:20.751459 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:21.748630 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:21.748595 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:21.748793 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:21.748644 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:21.748865 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:21.748764 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:21.748948 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:21.748902 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:22.749121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:22.749081 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:22.749655 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:22.749219 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:23.748789 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:23.748742 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:23.748944 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:23.748746 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:23.748944 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:23.748884 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-zmtbt" podUID="02351e99-a2a3-489f-b4dc-f8f12e779f0d" Apr 17 09:21:23.749043 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:23.748964 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:21:24.402547 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:24.402505 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:24.402950 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.402652 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:24.402950 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.402721 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:56.402706029 +0000 UTC m=+66.215619181 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 09:21:24.503893 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:24.503852 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:24.504039 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.504004 2581 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 17 09:21:24.504082 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.504070 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret podName:819cb802-41c1-426d-bcc4-c96157d9e5ed nodeName:}" failed. No retries permitted until 2026-04-17 09:21:56.504055053 +0000 UTC m=+66.316968205 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret") pod "global-pull-secret-syncer-hbnq9" (UID: "819cb802-41c1-426d-bcc4-c96157d9e5ed") : object "kube-system"/"original-pull-secret" not registered Apr 17 09:21:24.604955 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:24.604916 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:24.605143 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.605098 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 09:21:24.605143 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.605120 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 09:21:24.605143 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.605132 2581 projected.go:194] Error preparing data for projected volume kube-api-access-md7x7 for pod openshift-network-diagnostics/network-check-target-zmtbt: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:24.605276 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.605194 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7 podName:02351e99-a2a3-489f-b4dc-f8f12e779f0d nodeName:}" failed. No retries permitted until 2026-04-17 09:21:56.605175801 +0000 UTC m=+66.418088953 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-md7x7" (UniqueName: "kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7") pod "network-check-target-zmtbt" (UID: "02351e99-a2a3-489f-b4dc-f8f12e779f0d") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 09:21:24.748823 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:24.748744 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:24.748920 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:24.748896 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-hbnq9" podUID="819cb802-41c1-426d-bcc4-c96157d9e5ed" Apr 17 09:21:24.930448 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:24.930207 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerStarted","Data":"fff7a7ad048b79a2ed3067e3dd65b6e9a6583cd1444d7da40b807bd171628666"} Apr 17 09:21:25.568693 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.568661 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-129-70.ec2.internal" event="NodeReady" Apr 17 09:21:25.569100 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.568809 2581 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 17 09:21:25.598860 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.598834 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb"] Apr 17 09:21:25.615894 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.615413 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-5d4b47d6b8-zvhmm"] Apr 17 09:21:25.615894 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.615566 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.618873 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.618359 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"managed-serviceaccount-dockercfg-m4pbl\"" Apr 17 09:21:25.618873 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.618443 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"open-cluster-management-image-pull-credentials\"" Apr 17 09:21:25.618873 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.618449 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"open-cluster-management-agent-addon\"/\"kube-root-ca.crt\"" Apr 17 09:21:25.618873 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.618359 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"open-cluster-management-agent-addon\"/\"openshift-service-ca.crt\"" Apr 17 09:21:25.618873 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.618645 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"managed-serviceaccount-hub-kubeconfig\"" Apr 17 09:21:25.627845 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.627822 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4"] Apr 17 09:21:25.627999 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.627959 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.630317 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.630298 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 17 09:21:25.630429 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.630313 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 17 09:21:25.630429 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.630320 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-brmrj\"" Apr 17 09:21:25.630652 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.630335 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 17 09:21:25.635761 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.635732 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 17 09:21:25.643243 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.643212 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld"] Apr 17 09:21:25.643383 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.643365 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.646046 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.646026 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-open-cluster-management.io-proxy-agent-signer-client-cert\"" Apr 17 09:21:25.646046 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.646029 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-ca\"" Apr 17 09:21:25.646197 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.646047 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-service-proxy-server-certificates\"" Apr 17 09:21:25.646197 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.646113 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-hub-kubeconfig\"" Apr 17 09:21:25.655089 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.655068 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb"] Apr 17 09:21:25.655089 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.655091 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld"] Apr 17 09:21:25.655207 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.655099 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4"] Apr 17 09:21:25.655207 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.655106 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d4b47d6b8-zvhmm"] Apr 17 09:21:25.655207 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.655115 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-9mk5g"] Apr 17 09:21:25.655207 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.655190 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.657377 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.657359 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"work-manager-hub-kubeconfig\"" Apr 17 09:21:25.667380 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.667210 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9mk5g"] Apr 17 09:21:25.667380 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.667359 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:25.669661 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.669641 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 17 09:21:25.669780 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.669650 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 17 09:21:25.669865 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.669848 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 17 09:21:25.669926 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.669851 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9szrj\"" Apr 17 09:21:25.712326 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.712298 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds9jk\" (UniqueName: \"kubernetes.io/projected/a6e23fb9-b461-4f4b-85b4-6de703386fc0-kube-api-access-ds9jk\") pod \"managed-serviceaccount-addon-agent-74fff5df47-2wcfb\" (UID: \"a6e23fb9-b461-4f4b-85b4-6de703386fc0\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.712477 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.712342 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/a6e23fb9-b461-4f4b-85b4-6de703386fc0-hub-kubeconfig\") pod \"managed-serviceaccount-addon-agent-74fff5df47-2wcfb\" (UID: \"a6e23fb9-b461-4f4b-85b4-6de703386fc0\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.716450 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.716424 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bnzr4"] Apr 17 09:21:25.732302 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.732271 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bnzr4"] Apr 17 09:21:25.732425 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.732340 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:25.734650 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.734631 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 17 09:21:25.734802 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.734781 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-nblxl\"" Apr 17 09:21:25.734879 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.734833 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 17 09:21:25.748808 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.748787 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:25.748920 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.748790 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:25.751169 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.751148 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 17 09:21:25.751278 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.751217 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 17 09:21:25.751278 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.751242 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-8wjkc\"" Apr 17 09:21:25.751278 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.751242 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 17 09:21:25.751438 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.751264 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-4w67p\"" Apr 17 09:21:25.813431 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813399 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-image-registry-private-configuration\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.813431 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813436 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.813644 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813466 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/a6e23fb9-b461-4f4b-85b4-6de703386fc0-hub-kubeconfig\") pod \"managed-serviceaccount-addon-agent-74fff5df47-2wcfb\" (UID: \"a6e23fb9-b461-4f4b-85b4-6de703386fc0\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.813644 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813498 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkx5b\" (UniqueName: \"kubernetes.io/projected/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-kube-api-access-pkx5b\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.813644 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813526 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6t77\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-kube-api-access-q6t77\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.813644 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813545 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:25.813644 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813562 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.813644 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813584 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-hub\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.813892 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813633 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.813892 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813713 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-trusted-ca\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.813892 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813749 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-installation-pull-secrets\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.813892 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813805 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/499d1107-680a-4631-8788-202d1b1d65d5-tmp\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.813892 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813854 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-registry-certificates\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.813892 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813877 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-bound-sa-token\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.813892 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813894 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"klusterlet-config\" (UniqueName: \"kubernetes.io/secret/499d1107-680a-4631-8788-202d1b1d65d5-klusterlet-config\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.814179 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813920 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ds9jk\" (UniqueName: \"kubernetes.io/projected/a6e23fb9-b461-4f4b-85b4-6de703386fc0-kube-api-access-ds9jk\") pod \"managed-serviceaccount-addon-agent-74fff5df47-2wcfb\" (UID: \"a6e23fb9-b461-4f4b-85b4-6de703386fc0\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.814179 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813940 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-ca\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.814179 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813962 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2pr2\" (UniqueName: \"kubernetes.io/projected/696f7ce6-641f-473d-b364-c01310172e36-kube-api-access-b2pr2\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:25.814179 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.813977 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.814179 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.814005 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmk9z\" (UniqueName: \"kubernetes.io/projected/499d1107-680a-4631-8788-202d1b1d65d5-kube-api-access-hmk9z\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.814179 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.814063 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa5537a-c629-453b-a55f-7b99c548f514-ca-trust-extracted\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.817736 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.817720 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/a6e23fb9-b461-4f4b-85b4-6de703386fc0-hub-kubeconfig\") pod \"managed-serviceaccount-addon-agent-74fff5df47-2wcfb\" (UID: \"a6e23fb9-b461-4f4b-85b4-6de703386fc0\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.821964 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.821939 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds9jk\" (UniqueName: \"kubernetes.io/projected/a6e23fb9-b461-4f4b-85b4-6de703386fc0-kube-api-access-ds9jk\") pod \"managed-serviceaccount-addon-agent-74fff5df47-2wcfb\" (UID: \"a6e23fb9-b461-4f4b-85b4-6de703386fc0\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.915004 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.914967 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/499d1107-680a-4631-8788-202d1b1d65d5-tmp\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.915170 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915012 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-registry-certificates\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915170 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915032 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-bound-sa-token\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915170 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915049 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"klusterlet-config\" (UniqueName: \"kubernetes.io/secret/499d1107-680a-4631-8788-202d1b1d65d5-klusterlet-config\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.915381 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915200 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-ca\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.915381 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915253 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b2pr2\" (UniqueName: \"kubernetes.io/projected/696f7ce6-641f-473d-b364-c01310172e36-kube-api-access-b2pr2\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:25.915381 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915279 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.915381 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915327 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:25.915381 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915370 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hmk9z\" (UniqueName: \"kubernetes.io/projected/499d1107-680a-4631-8788-202d1b1d65d5-kube-api-access-hmk9z\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.915624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915398 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4061c0c-dd6a-4488-81f0-2b518afd6af0-config-volume\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:25.915624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915452 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa5537a-c629-453b-a55f-7b99c548f514-ca-trust-extracted\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915494 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-image-registry-private-configuration\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915492 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/499d1107-680a-4631-8788-202d1b1d65d5-tmp\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.915624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915557 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.915624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915597 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pkx5b\" (UniqueName: \"kubernetes.io/projected/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-kube-api-access-pkx5b\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915637 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-q6t77\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-kube-api-access-q6t77\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915661 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915679 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915700 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-hub\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915723 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915752 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-trusted-ca\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915763 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-registry-certificates\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915795 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c4061c0c-dd6a-4488-81f0-2b518afd6af0-tmp-dir\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915817 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jltk\" (UniqueName: \"kubernetes.io/projected/c4061c0c-dd6a-4488-81f0-2b518afd6af0-kube-api-access-4jltk\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:25.915919 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915845 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-installation-pull-secrets\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.916361 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.915926 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa5537a-c629-453b-a55f-7b99c548f514-ca-trust-extracted\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.916361 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:25.916046 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:25.916361 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:25.916117 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:26.416095368 +0000 UTC m=+36.229008535 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:21:25.916361 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:25.916219 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:21:25.916361 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:25.916234 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:21:25.916361 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.916260 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.916361 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:25.916289 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:26.4162737 +0000 UTC m=+36.229186866 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:21:25.917101 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.917082 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-trusted-ca\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.917832 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.917807 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"klusterlet-config\" (UniqueName: \"kubernetes.io/secret/499d1107-680a-4631-8788-202d1b1d65d5-klusterlet-config\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.918511 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.918492 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-installation-pull-secrets\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.918511 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.918509 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-image-registry-private-configuration\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.922883 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.922861 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-bound-sa-token\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.922883 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.922881 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2pr2\" (UniqueName: \"kubernetes.io/projected/696f7ce6-641f-473d-b364-c01310172e36-kube-api-access-b2pr2\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:25.926149 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.926104 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkx5b\" (UniqueName: \"kubernetes.io/projected/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-kube-api-access-pkx5b\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.926149 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.926114 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6t77\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-kube-api-access-q6t77\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:25.926731 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.926710 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmk9z\" (UniqueName: \"kubernetes.io/projected/499d1107-680a-4631-8788-202d1b1d65d5-kube-api-access-hmk9z\") pod \"klusterlet-addon-workmgr-6655d77c8c-569ld\" (UID: \"499d1107-680a-4631-8788-202d1b1d65d5\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:25.930617 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.930589 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.930699 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.930625 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-hub\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.930902 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.930882 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-ca\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.930943 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.930886 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/5a2a77c6-e47e-4ea2-abf8-01bea6fc9672-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-64b6b59bcb-5rbr4\" (UID: \"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.932089 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.932070 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" Apr 17 09:21:25.935064 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.935043 2581 generic.go:358] "Generic (PLEG): container finished" podID="bd655abe-79ad-487e-adbe-b874b344be93" containerID="fff7a7ad048b79a2ed3067e3dd65b6e9a6583cd1444d7da40b807bd171628666" exitCode=0 Apr 17 09:21:25.935140 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.935090 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerDied","Data":"fff7a7ad048b79a2ed3067e3dd65b6e9a6583cd1444d7da40b807bd171628666"} Apr 17 09:21:25.950406 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.950381 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:21:25.976359 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:25.976326 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:26.017320 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.017283 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.017463 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.017345 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4061c0c-dd6a-4488-81f0-2b518afd6af0-config-volume\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.017463 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.017436 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c4061c0c-dd6a-4488-81f0-2b518afd6af0-tmp-dir\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.017569 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.017463 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4jltk\" (UniqueName: \"kubernetes.io/projected/c4061c0c-dd6a-4488-81f0-2b518afd6af0-kube-api-access-4jltk\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.018184 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.018160 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:26.018280 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.018234 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:26.518213176 +0000 UTC m=+36.331126334 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:21:26.019202 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.019137 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c4061c0c-dd6a-4488-81f0-2b518afd6af0-tmp-dir\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.020102 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.020058 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4061c0c-dd6a-4488-81f0-2b518afd6af0-config-volume\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.033572 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.033508 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jltk\" (UniqueName: \"kubernetes.io/projected/c4061c0c-dd6a-4488-81f0-2b518afd6af0-kube-api-access-4jltk\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.137214 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.137185 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4"] Apr 17 09:21:26.140729 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:21:26.140705 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a2a77c6_e47e_4ea2_abf8_01bea6fc9672.slice/crio-d23fdd5caf55134a2e6b7b8b01457785065a3280193880384153e37a1967643b WatchSource:0}: Error finding container d23fdd5caf55134a2e6b7b8b01457785065a3280193880384153e37a1967643b: Status 404 returned error can't find the container with id d23fdd5caf55134a2e6b7b8b01457785065a3280193880384153e37a1967643b Apr 17 09:21:26.147546 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.147523 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb"] Apr 17 09:21:26.150238 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.150202 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld"] Apr 17 09:21:26.151981 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:21:26.151953 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6e23fb9_b461_4f4b_85b4_6de703386fc0.slice/crio-265530366b3df415c078d6d2a07189d326d531f51fbf00399502342a5836f496 WatchSource:0}: Error finding container 265530366b3df415c078d6d2a07189d326d531f51fbf00399502342a5836f496: Status 404 returned error can't find the container with id 265530366b3df415c078d6d2a07189d326d531f51fbf00399502342a5836f496 Apr 17 09:21:26.153801 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:21:26.153748 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod499d1107_680a_4631_8788_202d1b1d65d5.slice/crio-90f8b8dde3fe5f1b7bf174bfc3b11de3a71da2a639bb04ce7b37b4c2723c5806 WatchSource:0}: Error finding container 90f8b8dde3fe5f1b7bf174bfc3b11de3a71da2a639bb04ce7b37b4c2723c5806: Status 404 returned error can't find the container with id 90f8b8dde3fe5f1b7bf174bfc3b11de3a71da2a639bb04ce7b37b4c2723c5806 Apr 17 09:21:26.421131 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.421094 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:26.421131 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.421134 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:26.421585 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.421276 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:26.421585 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.421344 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:21:26.421585 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.421354 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:21:26.421585 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.421367 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:27.421348082 +0000 UTC m=+37.234261237 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:21:26.421585 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.421392 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:27.421380346 +0000 UTC m=+37.234293498 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:21:26.521997 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.521908 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:26.522149 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.522057 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:26.522149 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:26.522117 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:27.522102131 +0000 UTC m=+37.335015283 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:21:26.748559 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.748517 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:26.751211 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.751192 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 17 09:21:26.937828 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.937730 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" event={"ID":"499d1107-680a-4631-8788-202d1b1d65d5","Type":"ContainerStarted","Data":"90f8b8dde3fe5f1b7bf174bfc3b11de3a71da2a639bb04ce7b37b4c2723c5806"} Apr 17 09:21:26.938637 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.938615 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" event={"ID":"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672","Type":"ContainerStarted","Data":"d23fdd5caf55134a2e6b7b8b01457785065a3280193880384153e37a1967643b"} Apr 17 09:21:26.939539 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.939503 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" event={"ID":"a6e23fb9-b461-4f4b-85b4-6de703386fc0","Type":"ContainerStarted","Data":"265530366b3df415c078d6d2a07189d326d531f51fbf00399502342a5836f496"} Apr 17 09:21:26.941907 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.941886 2581 generic.go:358] "Generic (PLEG): container finished" podID="bd655abe-79ad-487e-adbe-b874b344be93" containerID="38d7804f7504bdb3d67dc12a5fda0b27efc88af2d5956399baa5d0da01ce110c" exitCode=0 Apr 17 09:21:26.941994 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:26.941917 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerDied","Data":"38d7804f7504bdb3d67dc12a5fda0b27efc88af2d5956399baa5d0da01ce110c"} Apr 17 09:21:27.431910 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:27.431867 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:27.432105 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:27.431929 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:27.432105 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:27.432084 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:27.432222 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:27.432133 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:21:27.432222 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:27.432146 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:21:27.432222 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:27.432160 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:29.432139861 +0000 UTC m=+39.245053020 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:21:27.432222 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:27.432194 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:29.432179185 +0000 UTC m=+39.245092339 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:21:27.533314 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:27.533175 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:27.533497 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:27.533328 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:27.533497 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:27.533394 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:29.533375498 +0000 UTC m=+39.346288654 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:21:27.948663 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:27.948623 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" event={"ID":"bd655abe-79ad-487e-adbe-b874b344be93","Type":"ContainerStarted","Data":"c19ce401d4de2ac92bbe0b29944b5daec7bc4fbc866c84efe71e0fa60b481266"} Apr 17 09:21:27.972165 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:27.972092 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2h6z9" podStartSLOduration=6.810775331 podStartE2EDuration="37.972074793s" podCreationTimestamp="2026-04-17 09:20:50 +0000 UTC" firstStartedPulling="2026-04-17 09:20:53.494387342 +0000 UTC m=+3.307300494" lastFinishedPulling="2026-04-17 09:21:24.655686789 +0000 UTC m=+34.468599956" observedRunningTime="2026-04-17 09:21:27.970521118 +0000 UTC m=+37.783434293" watchObservedRunningTime="2026-04-17 09:21:27.972074793 +0000 UTC m=+37.784987963" Apr 17 09:21:29.450185 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:29.450143 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:29.450624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:29.450288 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:29.450624 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:29.450300 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:21:29.450624 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:29.450327 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:21:29.450624 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:29.450395 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:29.450624 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:29.450403 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:33.450381307 +0000 UTC m=+43.263294463 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:21:29.450624 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:29.450445 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:33.450430632 +0000 UTC m=+43.263343786 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:21:29.551419 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:29.551341 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:29.551594 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:29.551524 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:29.551655 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:29.551609 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:33.551584451 +0000 UTC m=+43.364497611 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:21:31.958531 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:31.958331 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" event={"ID":"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672","Type":"ContainerStarted","Data":"df8733d17b53c0f724de0fbd9160c873ae97eb1136b053260853bd5ae3c8903e"} Apr 17 09:21:31.959732 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:31.959707 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" event={"ID":"a6e23fb9-b461-4f4b-85b4-6de703386fc0","Type":"ContainerStarted","Data":"cb096e3e9f090b1220e6cbfa1030192181a2dcc78e75aecd543e7b03468d3e9a"} Apr 17 09:21:31.975023 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:31.974979 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" podStartSLOduration=12.082520777 podStartE2EDuration="16.974965176s" podCreationTimestamp="2026-04-17 09:21:15 +0000 UTC" firstStartedPulling="2026-04-17 09:21:26.154444331 +0000 UTC m=+35.967357485" lastFinishedPulling="2026-04-17 09:21:31.046888714 +0000 UTC m=+40.859801884" observedRunningTime="2026-04-17 09:21:31.974242495 +0000 UTC m=+41.787155669" watchObservedRunningTime="2026-04-17 09:21:31.974965176 +0000 UTC m=+41.787878350" Apr 17 09:21:33.482365 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:33.482329 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:33.482365 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:33.482369 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:33.482846 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:33.482480 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:33.482846 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:33.482496 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:21:33.482846 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:33.482507 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:21:33.482846 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:33.482547 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:41.482531585 +0000 UTC m=+51.295444738 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:21:33.482846 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:33.482566 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:41.482557954 +0000 UTC m=+51.295471106 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:21:33.583474 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:33.583442 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:33.583618 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:33.583590 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:33.583661 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:33.583652 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:41.583637741 +0000 UTC m=+51.396550893 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:21:33.965426 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:33.965376 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" event={"ID":"499d1107-680a-4631-8788-202d1b1d65d5","Type":"ContainerStarted","Data":"82ea494d8524f2277f1a55cfd9c316817c4010985e143590fe8f6815fdd1e8f0"} Apr 17 09:21:33.965662 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:33.965628 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:33.967445 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:33.967424 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:21:33.980994 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:33.980954 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" podStartSLOduration=11.485838465 podStartE2EDuration="18.980941573s" podCreationTimestamp="2026-04-17 09:21:15 +0000 UTC" firstStartedPulling="2026-04-17 09:21:26.155581491 +0000 UTC m=+35.968494644" lastFinishedPulling="2026-04-17 09:21:33.650684599 +0000 UTC m=+43.463597752" observedRunningTime="2026-04-17 09:21:33.979950836 +0000 UTC m=+43.792864011" watchObservedRunningTime="2026-04-17 09:21:33.980941573 +0000 UTC m=+43.793854746" Apr 17 09:21:37.974413 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:37.974377 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" event={"ID":"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672","Type":"ContainerStarted","Data":"6f1f8a84b5747f598ca600a777b862a76f38377cb66dd6f9495bf6a9f6dbdd48"} Apr 17 09:21:37.974413 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:37.974414 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" event={"ID":"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672","Type":"ContainerStarted","Data":"4df08455d6641431e8e52e6630cccc0d5c38bda31bafde313953f3fc0473c632"} Apr 17 09:21:37.993355 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:37.993313 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" podStartSLOduration=12.043713869 podStartE2EDuration="22.99330073s" podCreationTimestamp="2026-04-17 09:21:15 +0000 UTC" firstStartedPulling="2026-04-17 09:21:26.142364745 +0000 UTC m=+35.955277897" lastFinishedPulling="2026-04-17 09:21:37.091951605 +0000 UTC m=+46.904864758" observedRunningTime="2026-04-17 09:21:37.991919899 +0000 UTC m=+47.804833065" watchObservedRunningTime="2026-04-17 09:21:37.99330073 +0000 UTC m=+47.806213925" Apr 17 09:21:41.543113 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:41.543081 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:41.543113 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:41.543117 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:41.543633 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:41.543218 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:21:41.543633 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:41.543230 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:21:41.543633 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:41.543229 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:41.543633 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:41.543309 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:57.54329408 +0000 UTC m=+67.356207233 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:21:41.543633 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:41.543323 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:57.543316716 +0000 UTC m=+67.356229870 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:21:41.643912 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:41.643874 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:41.644071 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:41.644030 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:41.644108 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:41.644101 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:21:57.644084808 +0000 UTC m=+67.456997960 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:21:49.927572 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:49.927535 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-64h7n" Apr 17 09:21:56.454102 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.454058 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:21:56.456756 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.456736 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 17 09:21:56.464731 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:56.464713 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 17 09:21:56.464796 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:56.464784 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:00.464753498 +0000 UTC m=+130.277666650 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : secret "metrics-daemon-secret" not found Apr 17 09:21:56.555230 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.555196 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:56.557881 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.557863 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 17 09:21:56.569347 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.569327 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/819cb802-41c1-426d-bcc4-c96157d9e5ed-original-pull-secret\") pod \"global-pull-secret-syncer-hbnq9\" (UID: \"819cb802-41c1-426d-bcc4-c96157d9e5ed\") " pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:56.656193 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.656153 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:56.658990 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.658971 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 17 09:21:56.669135 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.669112 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 17 09:21:56.680393 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.680366 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7x7\" (UniqueName: \"kubernetes.io/projected/02351e99-a2a3-489f-b4dc-f8f12e779f0d-kube-api-access-md7x7\") pod \"network-check-target-zmtbt\" (UID: \"02351e99-a2a3-489f-b4dc-f8f12e779f0d\") " pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:56.757581 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.757508 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-hbnq9" Apr 17 09:21:56.867511 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.867481 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-hbnq9"] Apr 17 09:21:56.871099 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:21:56.871075 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod819cb802_41c1_426d_bcc4_c96157d9e5ed.slice/crio-38d904879175e3768b2955ec501c9f40817b7999d27546a658108f1d9019ecb5 WatchSource:0}: Error finding container 38d904879175e3768b2955ec501c9f40817b7999d27546a658108f1d9019ecb5: Status 404 returned error can't find the container with id 38d904879175e3768b2955ec501c9f40817b7999d27546a658108f1d9019ecb5 Apr 17 09:21:56.960494 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.960465 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-4w67p\"" Apr 17 09:21:56.968562 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:56.968536 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:21:57.009468 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:57.009379 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-hbnq9" event={"ID":"819cb802-41c1-426d-bcc4-c96157d9e5ed","Type":"ContainerStarted","Data":"38d904879175e3768b2955ec501c9f40817b7999d27546a658108f1d9019ecb5"} Apr 17 09:21:57.077678 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:57.077568 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-zmtbt"] Apr 17 09:21:57.080104 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:21:57.080077 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02351e99_a2a3_489f_b4dc_f8f12e779f0d.slice/crio-b569923ac73ac0c969e8c61aec2737c8e81c87dce22d87a1309aaec9256641f4 WatchSource:0}: Error finding container b569923ac73ac0c969e8c61aec2737c8e81c87dce22d87a1309aaec9256641f4: Status 404 returned error can't find the container with id b569923ac73ac0c969e8c61aec2737c8e81c87dce22d87a1309aaec9256641f4 Apr 17 09:21:57.562780 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:57.562729 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:21:57.563185 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:57.562790 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:21:57.563185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:57.562878 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:21:57.563185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:57.562944 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:21:57.563185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:57.562957 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:29.56293798 +0000 UTC m=+99.375851138 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:21:57.563185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:57.562965 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:21:57.563185 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:57.563021 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:29.563002319 +0000 UTC m=+99.375915484 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:21:57.664175 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:57.664138 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:21:57.664361 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:57.664324 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:21:57.664430 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:21:57.664394 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:22:29.664379396 +0000 UTC m=+99.477292549 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:21:58.012416 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:21:58.012381 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-zmtbt" event={"ID":"02351e99-a2a3-489f-b4dc-f8f12e779f0d","Type":"ContainerStarted","Data":"b569923ac73ac0c969e8c61aec2737c8e81c87dce22d87a1309aaec9256641f4"} Apr 17 09:22:02.023273 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:02.023225 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-hbnq9" event={"ID":"819cb802-41c1-426d-bcc4-c96157d9e5ed","Type":"ContainerStarted","Data":"7427209faa800576231eaf7c6df38b799fd688f6317862326c7fa21bdad7a9e8"} Apr 17 09:22:02.024539 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:02.024501 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-zmtbt" event={"ID":"02351e99-a2a3-489f-b4dc-f8f12e779f0d","Type":"ContainerStarted","Data":"82a586f67ac01f9f51c5758f8536c36fe3a8f649ea5afd58e3480ce04038302f"} Apr 17 09:22:02.024679 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:02.024665 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:22:02.038131 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:02.038094 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-hbnq9" podStartSLOduration=66.558960098 podStartE2EDuration="1m11.038081796s" podCreationTimestamp="2026-04-17 09:20:51 +0000 UTC" firstStartedPulling="2026-04-17 09:21:56.872711143 +0000 UTC m=+66.685624309" lastFinishedPulling="2026-04-17 09:22:01.351832854 +0000 UTC m=+71.164746007" observedRunningTime="2026-04-17 09:22:02.037287102 +0000 UTC m=+71.850200273" watchObservedRunningTime="2026-04-17 09:22:02.038081796 +0000 UTC m=+71.850994968" Apr 17 09:22:02.052172 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:02.052128 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-zmtbt" podStartSLOduration=66.786487332 podStartE2EDuration="1m11.052113806s" podCreationTimestamp="2026-04-17 09:20:51 +0000 UTC" firstStartedPulling="2026-04-17 09:21:57.081964601 +0000 UTC m=+66.894877753" lastFinishedPulling="2026-04-17 09:22:01.347591071 +0000 UTC m=+71.160504227" observedRunningTime="2026-04-17 09:22:02.051714904 +0000 UTC m=+71.864628092" watchObservedRunningTime="2026-04-17 09:22:02.052113806 +0000 UTC m=+71.865026987" Apr 17 09:22:29.605797 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:29.605684 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:22:29.605797 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:29.605722 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:22:29.606211 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:22:29.605827 2581 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 17 09:22:29.606211 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:22:29.605838 2581 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-5d4b47d6b8-zvhmm: secret "image-registry-tls" not found Apr 17 09:22:29.606211 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:22:29.605839 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 09:22:29.606211 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:22:29.605898 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls podName:3aa5537a-c629-453b-a55f-7b99c548f514 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:33.605884009 +0000 UTC m=+163.418797166 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls") pod "image-registry-5d4b47d6b8-zvhmm" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514") : secret "image-registry-tls" not found Apr 17 09:22:29.606211 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:22:29.605910 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert podName:696f7ce6-641f-473d-b364-c01310172e36 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:33.60590435 +0000 UTC m=+163.418817502 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert") pod "ingress-canary-9mk5g" (UID: "696f7ce6-641f-473d-b364-c01310172e36") : secret "canary-serving-cert" not found Apr 17 09:22:29.706247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:29.706216 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:22:29.706381 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:22:29.706363 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 09:22:29.706437 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:22:29.706428 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls podName:c4061c0c-dd6a-4488-81f0-2b518afd6af0 nodeName:}" failed. No retries permitted until 2026-04-17 09:23:33.706413137 +0000 UTC m=+163.519326296 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls") pod "dns-default-bnzr4" (UID: "c4061c0c-dd6a-4488-81f0-2b518afd6af0") : secret "dns-default-metrics-tls" not found Apr 17 09:22:33.030406 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:22:33.030361 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-zmtbt" Apr 17 09:23:00.528007 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:00.527944 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:23:00.528589 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:23:00.528126 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 17 09:23:00.528589 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:23:00.528216 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs podName:504207a8-f3c9-45e3-a2cd-66458cf6eff8 nodeName:}" failed. No retries permitted until 2026-04-17 09:25:02.528193977 +0000 UTC m=+252.341107133 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs") pod "network-metrics-daemon-84m5n" (UID: "504207a8-f3c9-45e3-a2cd-66458cf6eff8") : secret "metrics-daemon-secret" not found Apr 17 09:23:03.205455 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:03.205426 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-6h6cx_56322e26-7d25-4937-af43-be08d98d2866/dns-node-resolver/0.log" Apr 17 09:23:04.405250 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:04.405221 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-sj6zp_9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31/node-ca/0.log" Apr 17 09:23:25.120547 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.120510 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-jmq8j"] Apr 17 09:23:25.123361 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.123345 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.125878 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.125856 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 17 09:23:25.126042 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.126026 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 17 09:23:25.126965 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.126947 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 17 09:23:25.127055 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.126984 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-fqvbv\"" Apr 17 09:23:25.127113 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.127062 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 17 09:23:25.131749 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.131728 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-jmq8j"] Apr 17 09:23:25.213162 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.213125 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e182d14d-d75c-42e5-851c-7797d94b6f39-data-volume\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.213334 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.213178 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e182d14d-d75c-42e5-851c-7797d94b6f39-crio-socket\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.213334 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.213199 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e182d14d-d75c-42e5-851c-7797d94b6f39-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.213334 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.213280 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e182d14d-d75c-42e5-851c-7797d94b6f39-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.213334 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.213325 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz2vg\" (UniqueName: \"kubernetes.io/projected/e182d14d-d75c-42e5-851c-7797d94b6f39-kube-api-access-dz2vg\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314069 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314022 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e182d14d-d75c-42e5-851c-7797d94b6f39-data-volume\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314084 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e182d14d-d75c-42e5-851c-7797d94b6f39-crio-socket\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314105 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e182d14d-d75c-42e5-851c-7797d94b6f39-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314138 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e182d14d-d75c-42e5-851c-7797d94b6f39-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314169 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dz2vg\" (UniqueName: \"kubernetes.io/projected/e182d14d-d75c-42e5-851c-7797d94b6f39-kube-api-access-dz2vg\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314171 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e182d14d-d75c-42e5-851c-7797d94b6f39-crio-socket\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314470 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314450 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e182d14d-d75c-42e5-851c-7797d94b6f39-data-volume\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.314642 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.314626 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e182d14d-d75c-42e5-851c-7797d94b6f39-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.316485 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.316462 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e182d14d-d75c-42e5-851c-7797d94b6f39-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.322448 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.322422 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz2vg\" (UniqueName: \"kubernetes.io/projected/e182d14d-d75c-42e5-851c-7797d94b6f39-kube-api-access-dz2vg\") pod \"insights-runtime-extractor-jmq8j\" (UID: \"e182d14d-d75c-42e5-851c-7797d94b6f39\") " pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.433391 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.433306 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-jmq8j" Apr 17 09:23:25.552898 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:25.552706 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-jmq8j"] Apr 17 09:23:25.555851 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:23:25.555822 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode182d14d_d75c_42e5_851c_7797d94b6f39.slice/crio-9e42b9d651edce5955d8d01fab52d82317848aa88a10f741520eac392265e19a WatchSource:0}: Error finding container 9e42b9d651edce5955d8d01fab52d82317848aa88a10f741520eac392265e19a: Status 404 returned error can't find the container with id 9e42b9d651edce5955d8d01fab52d82317848aa88a10f741520eac392265e19a Apr 17 09:23:26.222467 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:26.222430 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-jmq8j" event={"ID":"e182d14d-d75c-42e5-851c-7797d94b6f39","Type":"ContainerStarted","Data":"216cccc33b487f29b98ddba0b67d26bd3fde7c66b5b8e29aa26b35b1e23c57e7"} Apr 17 09:23:26.222467 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:26.222469 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-jmq8j" event={"ID":"e182d14d-d75c-42e5-851c-7797d94b6f39","Type":"ContainerStarted","Data":"9e42b9d651edce5955d8d01fab52d82317848aa88a10f741520eac392265e19a"} Apr 17 09:23:27.226903 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:27.226870 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-jmq8j" event={"ID":"e182d14d-d75c-42e5-851c-7797d94b6f39","Type":"ContainerStarted","Data":"bfd43cc04c2a340b47f3f2f7c01d757b43d8d6204750c67acdc4b86ecf372eef"} Apr 17 09:23:28.230852 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:28.230813 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-jmq8j" event={"ID":"e182d14d-d75c-42e5-851c-7797d94b6f39","Type":"ContainerStarted","Data":"16fb4e1071df79735699a492da8f853f13e3bfdc577e25170aa725d07a1f707a"} Apr 17 09:23:28.251850 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:28.251796 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-jmq8j" podStartSLOduration=1.206992337 podStartE2EDuration="3.251763207s" podCreationTimestamp="2026-04-17 09:23:25 +0000 UTC" firstStartedPulling="2026-04-17 09:23:25.613188658 +0000 UTC m=+155.426101825" lastFinishedPulling="2026-04-17 09:23:27.657959537 +0000 UTC m=+157.470872695" observedRunningTime="2026-04-17 09:23:28.250634285 +0000 UTC m=+158.063547497" watchObservedRunningTime="2026-04-17 09:23:28.251763207 +0000 UTC m=+158.064676404" Apr 17 09:23:28.641749 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:23:28.641648 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[registry-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" podUID="3aa5537a-c629-453b-a55f-7b99c548f514" Apr 17 09:23:28.685900 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:23:28.685859 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-9mk5g" podUID="696f7ce6-641f-473d-b364-c01310172e36" Apr 17 09:23:28.741205 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:23:28.741166 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-bnzr4" podUID="c4061c0c-dd6a-4488-81f0-2b518afd6af0" Apr 17 09:23:28.761945 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:23:28.761909 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-certs], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-multus/network-metrics-daemon-84m5n" podUID="504207a8-f3c9-45e3-a2cd-66458cf6eff8" Apr 17 09:23:29.233025 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:29.232998 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bnzr4" Apr 17 09:23:29.233330 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:29.232999 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:23:32.243737 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:32.243705 2581 generic.go:358] "Generic (PLEG): container finished" podID="a6e23fb9-b461-4f4b-85b4-6de703386fc0" containerID="cb096e3e9f090b1220e6cbfa1030192181a2dcc78e75aecd543e7b03468d3e9a" exitCode=255 Apr 17 09:23:32.244136 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:32.243789 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" event={"ID":"a6e23fb9-b461-4f4b-85b4-6de703386fc0","Type":"ContainerDied","Data":"cb096e3e9f090b1220e6cbfa1030192181a2dcc78e75aecd543e7b03468d3e9a"} Apr 17 09:23:32.244136 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:32.244083 2581 scope.go:117] "RemoveContainer" containerID="cb096e3e9f090b1220e6cbfa1030192181a2dcc78e75aecd543e7b03468d3e9a" Apr 17 09:23:33.247934 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.247896 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-74fff5df47-2wcfb" event={"ID":"a6e23fb9-b461-4f4b-85b4-6de703386fc0","Type":"ContainerStarted","Data":"386673b83253a396d7351f059d4ac158a05d015f674bf0c44714de3b3538a595"} Apr 17 09:23:33.687030 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.686922 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:23:33.687030 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.686977 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:23:33.689413 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.689386 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"image-registry-5d4b47d6b8-zvhmm\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:23:33.689523 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.689469 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/696f7ce6-641f-473d-b364-c01310172e36-cert\") pod \"ingress-canary-9mk5g\" (UID: \"696f7ce6-641f-473d-b364-c01310172e36\") " pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:23:33.736889 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.736859 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-brmrj\"" Apr 17 09:23:33.744851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.744826 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:23:33.788227 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.788196 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:23:33.790848 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.790788 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4061c0c-dd6a-4488-81f0-2b518afd6af0-metrics-tls\") pod \"dns-default-bnzr4\" (UID: \"c4061c0c-dd6a-4488-81f0-2b518afd6af0\") " pod="openshift-dns/dns-default-bnzr4" Apr 17 09:23:33.866263 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.866237 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d4b47d6b8-zvhmm"] Apr 17 09:23:33.868763 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:23:33.868737 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3aa5537a_c629_453b_a55f_7b99c548f514.slice/crio-4d078b4b179c53f289c175596a47cf644a3c31833231e2dbb814289b71e2cc68 WatchSource:0}: Error finding container 4d078b4b179c53f289c175596a47cf644a3c31833231e2dbb814289b71e2cc68: Status 404 returned error can't find the container with id 4d078b4b179c53f289c175596a47cf644a3c31833231e2dbb814289b71e2cc68 Apr 17 09:23:33.966498 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:33.966432 2581 prober.go:120] "Probe failed" probeType="Readiness" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" podUID="499d1107-680a-4631-8788-202d1b1d65d5" containerName="acm-agent" probeResult="failure" output="Get \"http://10.132.0.9:8000/readyz\": dial tcp 10.132.0.9:8000: connect: connection refused" Apr 17 09:23:34.035934 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.035905 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-nblxl\"" Apr 17 09:23:34.043928 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.043902 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bnzr4" Apr 17 09:23:34.158621 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.158586 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bnzr4"] Apr 17 09:23:34.161827 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:23:34.161800 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4061c0c_dd6a_4488_81f0_2b518afd6af0.slice/crio-a3ec2e38c931b5e72259cf97d318b860da4df58c8b9361d7d6a2b7c16149fe72 WatchSource:0}: Error finding container a3ec2e38c931b5e72259cf97d318b860da4df58c8b9361d7d6a2b7c16149fe72: Status 404 returned error can't find the container with id a3ec2e38c931b5e72259cf97d318b860da4df58c8b9361d7d6a2b7c16149fe72 Apr 17 09:23:34.251517 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.251434 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bnzr4" event={"ID":"c4061c0c-dd6a-4488-81f0-2b518afd6af0","Type":"ContainerStarted","Data":"a3ec2e38c931b5e72259cf97d318b860da4df58c8b9361d7d6a2b7c16149fe72"} Apr 17 09:23:34.252794 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.252753 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" event={"ID":"3aa5537a-c629-453b-a55f-7b99c548f514","Type":"ContainerStarted","Data":"02a8a3741ef07182d158b6d521d25495e78bb632912e0c5d5a8ef6c6e1a3754d"} Apr 17 09:23:34.252895 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.252802 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" event={"ID":"3aa5537a-c629-453b-a55f-7b99c548f514","Type":"ContainerStarted","Data":"4d078b4b179c53f289c175596a47cf644a3c31833231e2dbb814289b71e2cc68"} Apr 17 09:23:34.253011 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.252992 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:23:34.254090 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.254069 2581 generic.go:358] "Generic (PLEG): container finished" podID="499d1107-680a-4631-8788-202d1b1d65d5" containerID="82ea494d8524f2277f1a55cfd9c316817c4010985e143590fe8f6815fdd1e8f0" exitCode=1 Apr 17 09:23:34.254186 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.254106 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" event={"ID":"499d1107-680a-4631-8788-202d1b1d65d5","Type":"ContainerDied","Data":"82ea494d8524f2277f1a55cfd9c316817c4010985e143590fe8f6815fdd1e8f0"} Apr 17 09:23:34.254361 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.254350 2581 scope.go:117] "RemoveContainer" containerID="82ea494d8524f2277f1a55cfd9c316817c4010985e143590fe8f6815fdd1e8f0" Apr 17 09:23:34.271181 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:34.271132 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" podStartSLOduration=163.271114673 podStartE2EDuration="2m43.271114673s" podCreationTimestamp="2026-04-17 09:20:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:23:34.270395188 +0000 UTC m=+164.083308362" watchObservedRunningTime="2026-04-17 09:23:34.271114673 +0000 UTC m=+164.084027849" Apr 17 09:23:35.257965 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:35.257928 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" event={"ID":"499d1107-680a-4631-8788-202d1b1d65d5","Type":"ContainerStarted","Data":"a1dde9caf26e39d5a3dee360cf4279a9ac47db14ab12c34fb1b88252ff327167"} Apr 17 09:23:35.258441 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:35.258338 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:23:35.258930 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:35.258912 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-6655d77c8c-569ld" Apr 17 09:23:36.262247 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:36.262207 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bnzr4" event={"ID":"c4061c0c-dd6a-4488-81f0-2b518afd6af0","Type":"ContainerStarted","Data":"9f3d34169d78dca7f48b621184b6a743fc064e50eee69eb30c25dd4b04260f51"} Apr 17 09:23:36.262622 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:36.262255 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bnzr4" event={"ID":"c4061c0c-dd6a-4488-81f0-2b518afd6af0","Type":"ContainerStarted","Data":"5fd6e6493dd43c53dd9b9cb3cafe1ce055e73a189da90fbb78b05ee4870f52da"} Apr 17 09:23:36.278251 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:36.278199 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bnzr4" podStartSLOduration=129.55701067 podStartE2EDuration="2m11.278184878s" podCreationTimestamp="2026-04-17 09:21:25 +0000 UTC" firstStartedPulling="2026-04-17 09:23:34.163731311 +0000 UTC m=+163.976644464" lastFinishedPulling="2026-04-17 09:23:35.884905515 +0000 UTC m=+165.697818672" observedRunningTime="2026-04-17 09:23:36.277704352 +0000 UTC m=+166.090617527" watchObservedRunningTime="2026-04-17 09:23:36.278184878 +0000 UTC m=+166.091098053" Apr 17 09:23:37.265308 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:37.265279 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-bnzr4" Apr 17 09:23:38.891210 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.891178 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-h96rp"] Apr 17 09:23:38.894550 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.894528 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:38.897099 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.897075 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 17 09:23:38.897203 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.897183 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 17 09:23:38.897309 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.897296 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 17 09:23:38.898391 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.898370 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 17 09:23:38.898391 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.898386 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 17 09:23:38.898503 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.898446 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-hwqkh\"" Apr 17 09:23:38.898555 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:38.898516 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 17 09:23:39.033026 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.032983 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f29v\" (UniqueName: \"kubernetes.io/projected/02ffc515-c207-4767-b17d-e7944d4310b3-kube-api-access-4f29v\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033211 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033051 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-sys\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033211 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033074 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-wtmp\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033211 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033107 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-textfile\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033211 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033190 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-tls\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033367 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033226 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/02ffc515-c207-4767-b17d-e7944d4310b3-metrics-client-ca\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033367 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033282 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-accelerators-collector-config\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033367 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033310 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-root\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.033367 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.033341 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134407 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134366 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-sys\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134407 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134402 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-wtmp\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134422 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-textfile\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134443 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-tls\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134463 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/02ffc515-c207-4767-b17d-e7944d4310b3-metrics-client-ca\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134490 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-sys\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134506 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-accelerators-collector-config\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134562 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-root\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134573 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-wtmp\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.134624 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134600 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/02ffc515-c207-4767-b17d-e7944d4310b3-root\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.135022 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134706 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.135022 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134759 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4f29v\" (UniqueName: \"kubernetes.io/projected/02ffc515-c207-4767-b17d-e7944d4310b3-kube-api-access-4f29v\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.135022 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.134836 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-textfile\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.135173 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.135137 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-accelerators-collector-config\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.135236 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.135212 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/02ffc515-c207-4767-b17d-e7944d4310b3-metrics-client-ca\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.137009 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.136982 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.137094 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.137060 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/02ffc515-c207-4767-b17d-e7944d4310b3-node-exporter-tls\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.142679 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.142626 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f29v\" (UniqueName: \"kubernetes.io/projected/02ffc515-c207-4767-b17d-e7944d4310b3-kube-api-access-4f29v\") pod \"node-exporter-h96rp\" (UID: \"02ffc515-c207-4767-b17d-e7944d4310b3\") " pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.204455 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.204416 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-h96rp" Apr 17 09:23:39.212299 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:23:39.212267 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02ffc515_c207_4767_b17d_e7944d4310b3.slice/crio-fbc4e30632346eca1a077fa889f6aa75f4c687df151d456972d2d535168fe827 WatchSource:0}: Error finding container fbc4e30632346eca1a077fa889f6aa75f4c687df151d456972d2d535168fe827: Status 404 returned error can't find the container with id fbc4e30632346eca1a077fa889f6aa75f4c687df151d456972d2d535168fe827 Apr 17 09:23:39.272973 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.272934 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h96rp" event={"ID":"02ffc515-c207-4767-b17d-e7944d4310b3","Type":"ContainerStarted","Data":"fbc4e30632346eca1a077fa889f6aa75f4c687df151d456972d2d535168fe827"} Apr 17 09:23:39.748595 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:39.748556 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:23:40.276687 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:40.276597 2581 generic.go:358] "Generic (PLEG): container finished" podID="02ffc515-c207-4767-b17d-e7944d4310b3" containerID="1d8a7bd8f851dd01e76a59970a4e0e1f7b8f0e731c37aa7021d242beae07e30b" exitCode=0 Apr 17 09:23:40.276687 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:40.276658 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h96rp" event={"ID":"02ffc515-c207-4767-b17d-e7944d4310b3","Type":"ContainerDied","Data":"1d8a7bd8f851dd01e76a59970a4e0e1f7b8f0e731c37aa7021d242beae07e30b"} Apr 17 09:23:40.749914 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:40.749876 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:23:40.752605 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:40.752584 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9szrj\"" Apr 17 09:23:40.760692 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:40.760671 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9mk5g" Apr 17 09:23:40.876114 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:40.876080 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9mk5g"] Apr 17 09:23:40.878955 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:23:40.878930 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod696f7ce6_641f_473d_b364_c01310172e36.slice/crio-17255017bc719ae9c650ab49af344bff02f0b00ddffbe13347c9913089aaef40 WatchSource:0}: Error finding container 17255017bc719ae9c650ab49af344bff02f0b00ddffbe13347c9913089aaef40: Status 404 returned error can't find the container with id 17255017bc719ae9c650ab49af344bff02f0b00ddffbe13347c9913089aaef40 Apr 17 09:23:41.282414 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:41.282373 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h96rp" event={"ID":"02ffc515-c207-4767-b17d-e7944d4310b3","Type":"ContainerStarted","Data":"f1a2b26573844928bd5423aafc639cb4f880eb9a2639086936ec2a81b04d979f"} Apr 17 09:23:41.282414 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:41.282420 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h96rp" event={"ID":"02ffc515-c207-4767-b17d-e7944d4310b3","Type":"ContainerStarted","Data":"559ae394818e89042ab97e23daa6c9992e75f48ba1635e1e3c73beb0fc37a0ea"} Apr 17 09:23:41.283510 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:41.283486 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9mk5g" event={"ID":"696f7ce6-641f-473d-b364-c01310172e36","Type":"ContainerStarted","Data":"17255017bc719ae9c650ab49af344bff02f0b00ddffbe13347c9913089aaef40"} Apr 17 09:23:41.301263 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:41.301190 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-h96rp" podStartSLOduration=2.510341667 podStartE2EDuration="3.301176974s" podCreationTimestamp="2026-04-17 09:23:38 +0000 UTC" firstStartedPulling="2026-04-17 09:23:39.21411759 +0000 UTC m=+169.027030746" lastFinishedPulling="2026-04-17 09:23:40.004952898 +0000 UTC m=+169.817866053" observedRunningTime="2026-04-17 09:23:41.300085739 +0000 UTC m=+171.112998927" watchObservedRunningTime="2026-04-17 09:23:41.301176974 +0000 UTC m=+171.114090139" Apr 17 09:23:43.290170 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:43.290132 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9mk5g" event={"ID":"696f7ce6-641f-473d-b364-c01310172e36","Type":"ContainerStarted","Data":"e226a14aa75beaf1baab3867c0aac6130435129fef4f1294932a3b028b26f5f7"} Apr 17 09:23:43.305079 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:43.305018 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-9mk5g" podStartSLOduration=136.567899956 podStartE2EDuration="2m18.30500275s" podCreationTimestamp="2026-04-17 09:21:25 +0000 UTC" firstStartedPulling="2026-04-17 09:23:40.880905886 +0000 UTC m=+170.693819038" lastFinishedPulling="2026-04-17 09:23:42.61800868 +0000 UTC m=+172.430921832" observedRunningTime="2026-04-17 09:23:43.304249346 +0000 UTC m=+173.117162521" watchObservedRunningTime="2026-04-17 09:23:43.30500275 +0000 UTC m=+173.117915926" Apr 17 09:23:45.952039 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:45.951993 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" podUID="5a2a77c6-e47e-4ea2-abf8-01bea6fc9672" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 17 09:23:47.271881 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:47.271848 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bnzr4" Apr 17 09:23:53.749410 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:53.749376 2581 patch_prober.go:28] interesting pod/image-registry-5d4b47d6b8-zvhmm container/registry namespace/openshift-image-registry: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 17 09:23:53.749762 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:53.749428 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" podUID="3aa5537a-c629-453b-a55f-7b99c548f514" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 17 09:23:55.262213 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:55.262186 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:23:55.952391 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:55.952351 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" podUID="5a2a77c6-e47e-4ea2-abf8-01bea6fc9672" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 17 09:23:57.143564 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:23:57.143530 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-5d4b47d6b8-zvhmm"] Apr 17 09:24:05.951366 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:05.951314 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" podUID="5a2a77c6-e47e-4ea2-abf8-01bea6fc9672" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 17 09:24:05.951844 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:05.951408 2581 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" Apr 17 09:24:05.951929 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:05.951889 2581 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="service-proxy" containerStatusID={"Type":"cri-o","ID":"6f1f8a84b5747f598ca600a777b862a76f38377cb66dd6f9495bf6a9f6dbdd48"} pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" containerMessage="Container service-proxy failed liveness probe, will be restarted" Apr 17 09:24:05.951967 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:05.951951 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" podUID="5a2a77c6-e47e-4ea2-abf8-01bea6fc9672" containerName="service-proxy" containerID="cri-o://6f1f8a84b5747f598ca600a777b862a76f38377cb66dd6f9495bf6a9f6dbdd48" gracePeriod=30 Apr 17 09:24:06.349230 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:06.349141 2581 generic.go:358] "Generic (PLEG): container finished" podID="5a2a77c6-e47e-4ea2-abf8-01bea6fc9672" containerID="6f1f8a84b5747f598ca600a777b862a76f38377cb66dd6f9495bf6a9f6dbdd48" exitCode=2 Apr 17 09:24:06.349230 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:06.349189 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" event={"ID":"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672","Type":"ContainerDied","Data":"6f1f8a84b5747f598ca600a777b862a76f38377cb66dd6f9495bf6a9f6dbdd48"} Apr 17 09:24:06.349230 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:06.349214 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-64b6b59bcb-5rbr4" event={"ID":"5a2a77c6-e47e-4ea2-abf8-01bea6fc9672","Type":"ContainerStarted","Data":"7bf074b3c3e2324cff4b02a686c46d2e33e4b3764fa7826b342a6505a70ae43a"} Apr 17 09:24:22.162155 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.162093 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" podUID="3aa5537a-c629-453b-a55f-7b99c548f514" containerName="registry" containerID="cri-o://02a8a3741ef07182d158b6d521d25495e78bb632912e0c5d5a8ef6c6e1a3754d" gracePeriod=30 Apr 17 09:24:22.388268 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.388240 2581 generic.go:358] "Generic (PLEG): container finished" podID="3aa5537a-c629-453b-a55f-7b99c548f514" containerID="02a8a3741ef07182d158b6d521d25495e78bb632912e0c5d5a8ef6c6e1a3754d" exitCode=0 Apr 17 09:24:22.388392 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.388326 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" event={"ID":"3aa5537a-c629-453b-a55f-7b99c548f514","Type":"ContainerDied","Data":"02a8a3741ef07182d158b6d521d25495e78bb632912e0c5d5a8ef6c6e1a3754d"} Apr 17 09:24:22.388392 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.388376 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" event={"ID":"3aa5537a-c629-453b-a55f-7b99c548f514","Type":"ContainerDied","Data":"4d078b4b179c53f289c175596a47cf644a3c31833231e2dbb814289b71e2cc68"} Apr 17 09:24:22.388470 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.388391 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d078b4b179c53f289c175596a47cf644a3c31833231e2dbb814289b71e2cc68" Apr 17 09:24:22.396335 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.396317 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:24:22.457928 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.457897 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458071 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.457995 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-installation-pull-secrets\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458071 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458024 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa5537a-c629-453b-a55f-7b99c548f514-ca-trust-extracted\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458071 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458054 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-registry-certificates\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458233 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458088 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-image-registry-private-configuration\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458233 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458113 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-bound-sa-token\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458233 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458160 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6t77\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-kube-api-access-q6t77\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458233 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458188 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-trusted-ca\") pod \"3aa5537a-c629-453b-a55f-7b99c548f514\" (UID: \"3aa5537a-c629-453b-a55f-7b99c548f514\") " Apr 17 09:24:22.458743 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458659 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:24:22.458922 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.458765 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 09:24:22.460400 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.460369 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:24:22.460735 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.460700 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:24:22.460876 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.460719 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:24:22.460876 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.460791 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 09:24:22.460968 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.460953 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-kube-api-access-q6t77" (OuterVolumeSpecName: "kube-api-access-q6t77") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "kube-api-access-q6t77". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:24:22.469144 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.469115 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aa5537a-c629-453b-a55f-7b99c548f514-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "3aa5537a-c629-453b-a55f-7b99c548f514" (UID: "3aa5537a-c629-453b-a55f-7b99c548f514"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:24:22.559533 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559497 2581 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-installation-pull-secrets\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:22.559533 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559528 2581 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa5537a-c629-453b-a55f-7b99c548f514-ca-trust-extracted\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:22.559533 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559538 2581 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-registry-certificates\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:22.559765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559548 2581 reconciler_common.go:299] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/3aa5537a-c629-453b-a55f-7b99c548f514-image-registry-private-configuration\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:22.559765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559559 2581 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-bound-sa-token\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:22.559765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559567 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q6t77\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-kube-api-access-q6t77\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:22.559765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559578 2581 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa5537a-c629-453b-a55f-7b99c548f514-trusted-ca\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:22.559765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:22.559587 2581 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa5537a-c629-453b-a55f-7b99c548f514-registry-tls\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:24:23.390983 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:23.390936 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d4b47d6b8-zvhmm" Apr 17 09:24:23.406626 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:23.406595 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-5d4b47d6b8-zvhmm"] Apr 17 09:24:23.410999 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:23.410970 2581 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-5d4b47d6b8-zvhmm"] Apr 17 09:24:24.752555 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:24.752521 2581 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aa5537a-c629-453b-a55f-7b99c548f514" path="/var/lib/kubelet/pods/3aa5537a-c629-453b-a55f-7b99c548f514/volumes" Apr 17 09:24:29.711210 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:29.711186 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-h96rp_02ffc515-c207-4767-b17d-e7944d4310b3/init-textfile/0.log" Apr 17 09:24:29.911949 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:29.911914 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-h96rp_02ffc515-c207-4767-b17d-e7944d4310b3/node-exporter/0.log" Apr 17 09:24:30.110756 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:24:30.110677 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-h96rp_02ffc515-c207-4767-b17d-e7944d4310b3/kube-rbac-proxy/0.log" Apr 17 09:25:02.551554 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:02.551506 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:25:02.553844 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:02.553823 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/504207a8-f3c9-45e3-a2cd-66458cf6eff8-metrics-certs\") pod \"network-metrics-daemon-84m5n\" (UID: \"504207a8-f3c9-45e3-a2cd-66458cf6eff8\") " pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:25:02.852456 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:02.852372 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-8wjkc\"" Apr 17 09:25:02.860424 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:02.860397 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-84m5n" Apr 17 09:25:02.992121 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:02.992089 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-84m5n"] Apr 17 09:25:02.995277 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:25:02.995251 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod504207a8_f3c9_45e3_a2cd_66458cf6eff8.slice/crio-f5ec43e07c63e64ebe58667cd0ef9b457f6fbb484b8374805195d2bc5596c8b2 WatchSource:0}: Error finding container f5ec43e07c63e64ebe58667cd0ef9b457f6fbb484b8374805195d2bc5596c8b2: Status 404 returned error can't find the container with id f5ec43e07c63e64ebe58667cd0ef9b457f6fbb484b8374805195d2bc5596c8b2 Apr 17 09:25:03.497552 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:03.497517 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-84m5n" event={"ID":"504207a8-f3c9-45e3-a2cd-66458cf6eff8","Type":"ContainerStarted","Data":"f5ec43e07c63e64ebe58667cd0ef9b457f6fbb484b8374805195d2bc5596c8b2"} Apr 17 09:25:04.501633 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:04.501595 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-84m5n" event={"ID":"504207a8-f3c9-45e3-a2cd-66458cf6eff8","Type":"ContainerStarted","Data":"425d7f73768ddc753ffa350a903ccef0baee420c0fa84fe969609c1ba839d918"} Apr 17 09:25:04.502049 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:04.501644 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-84m5n" event={"ID":"504207a8-f3c9-45e3-a2cd-66458cf6eff8","Type":"ContainerStarted","Data":"ab2ff46570c3a49e9e3f0618a197e8ed549af1732b267626e0b35489a9ab4896"} Apr 17 09:25:04.517585 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:25:04.517537 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-84m5n" podStartSLOduration=253.407927355 podStartE2EDuration="4m14.517524287s" podCreationTimestamp="2026-04-17 09:20:50 +0000 UTC" firstStartedPulling="2026-04-17 09:25:02.997159289 +0000 UTC m=+252.810072457" lastFinishedPulling="2026-04-17 09:25:04.106756236 +0000 UTC m=+253.919669389" observedRunningTime="2026-04-17 09:25:04.51648789 +0000 UTC m=+254.329401065" watchObservedRunningTime="2026-04-17 09:25:04.517524287 +0000 UTC m=+254.330437463" Apr 17 09:27:49.120395 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.120352 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-6k5t6"] Apr 17 09:27:49.120954 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.120673 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3aa5537a-c629-453b-a55f-7b99c548f514" containerName="registry" Apr 17 09:27:49.120954 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.120690 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa5537a-c629-453b-a55f-7b99c548f514" containerName="registry" Apr 17 09:27:49.120954 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.120749 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="3aa5537a-c629-453b-a55f-7b99c548f514" containerName="registry" Apr 17 09:27:49.123480 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.123460 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.125720 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.125697 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-w9mvd\"" Apr 17 09:27:49.125848 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.125700 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Apr 17 09:27:49.126665 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.126647 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Apr 17 09:27:49.133008 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.132988 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-6k5t6"] Apr 17 09:27:49.284022 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.283979 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1f678af-01b3-4c77-9d81-0bf697091330-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-6k5t6\" (UID: \"b1f678af-01b3-4c77-9d81-0bf697091330\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.284200 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.284034 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzc4z\" (UniqueName: \"kubernetes.io/projected/b1f678af-01b3-4c77-9d81-0bf697091330-kube-api-access-xzc4z\") pod \"cert-manager-cainjector-8966b78d4-6k5t6\" (UID: \"b1f678af-01b3-4c77-9d81-0bf697091330\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.385343 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.385268 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1f678af-01b3-4c77-9d81-0bf697091330-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-6k5t6\" (UID: \"b1f678af-01b3-4c77-9d81-0bf697091330\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.385343 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.385310 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xzc4z\" (UniqueName: \"kubernetes.io/projected/b1f678af-01b3-4c77-9d81-0bf697091330-kube-api-access-xzc4z\") pod \"cert-manager-cainjector-8966b78d4-6k5t6\" (UID: \"b1f678af-01b3-4c77-9d81-0bf697091330\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.392905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.392877 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1f678af-01b3-4c77-9d81-0bf697091330-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-6k5t6\" (UID: \"b1f678af-01b3-4c77-9d81-0bf697091330\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.393110 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.393088 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzc4z\" (UniqueName: \"kubernetes.io/projected/b1f678af-01b3-4c77-9d81-0bf697091330-kube-api-access-xzc4z\") pod \"cert-manager-cainjector-8966b78d4-6k5t6\" (UID: \"b1f678af-01b3-4c77-9d81-0bf697091330\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.432430 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.432386 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" Apr 17 09:27:49.546958 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.546933 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-6k5t6"] Apr 17 09:27:49.549211 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:27:49.549188 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1f678af_01b3_4c77_9d81_0bf697091330.slice/crio-1fc39f9c8eeecef11d1a5d9886682bc3e32c5d50698fc119919dbee40e0759f3 WatchSource:0}: Error finding container 1fc39f9c8eeecef11d1a5d9886682bc3e32c5d50698fc119919dbee40e0759f3: Status 404 returned error can't find the container with id 1fc39f9c8eeecef11d1a5d9886682bc3e32c5d50698fc119919dbee40e0759f3 Apr 17 09:27:49.550924 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.550905 2581 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 09:27:49.929868 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:49.929837 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" event={"ID":"b1f678af-01b3-4c77-9d81-0bf697091330","Type":"ContainerStarted","Data":"1fc39f9c8eeecef11d1a5d9886682bc3e32c5d50698fc119919dbee40e0759f3"} Apr 17 09:27:53.941890 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:53.941855 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" event={"ID":"b1f678af-01b3-4c77-9d81-0bf697091330","Type":"ContainerStarted","Data":"dfaf9b564418ac1e2c03e891ffd4f8eb3fcb0dcd107aea98d25be26be67852ba"} Apr 17 09:27:53.955824 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:27:53.955752 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-8966b78d4-6k5t6" podStartSLOduration=1.590575883 podStartE2EDuration="4.955737257s" podCreationTimestamp="2026-04-17 09:27:49 +0000 UTC" firstStartedPulling="2026-04-17 09:27:49.551030578 +0000 UTC m=+419.363943730" lastFinishedPulling="2026-04-17 09:27:52.916191941 +0000 UTC m=+422.729105104" observedRunningTime="2026-04-17 09:27:53.955579648 +0000 UTC m=+423.768492822" watchObservedRunningTime="2026-04-17 09:27:53.955737257 +0000 UTC m=+423.768650431" Apr 17 09:28:47.183364 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.183312 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl"] Apr 17 09:28:47.186263 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.186248 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.188830 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.188804 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"kubeflow-trainer-config\"" Apr 17 09:28:47.188962 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.188845 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"kubeflow-trainer-webhook-cert\"" Apr 17 09:28:47.188962 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.188883 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"kube-root-ca.crt\"" Apr 17 09:28:47.189884 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.189867 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"kubeflow-trainer-controller-manager-dockercfg-96gqx\"" Apr 17 09:28:47.189970 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.189880 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"openshift-service-ca.crt\"" Apr 17 09:28:47.196133 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.196113 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl"] Apr 17 09:28:47.279525 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.279486 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl2kh\" (UniqueName: \"kubernetes.io/projected/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-kube-api-access-tl2kh\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.279689 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.279534 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeflow-trainer-config\" (UniqueName: \"kubernetes.io/configmap/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-kubeflow-trainer-config\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.279689 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.279602 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-cert\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.380610 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.380568 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubeflow-trainer-config\" (UniqueName: \"kubernetes.io/configmap/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-kubeflow-trainer-config\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.380765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.380624 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-cert\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.380765 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.380661 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tl2kh\" (UniqueName: \"kubernetes.io/projected/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-kube-api-access-tl2kh\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.381359 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.381329 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubeflow-trainer-config\" (UniqueName: \"kubernetes.io/configmap/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-kubeflow-trainer-config\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.383050 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.383028 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-cert\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.388101 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.388078 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl2kh\" (UniqueName: \"kubernetes.io/projected/62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8-kube-api-access-tl2kh\") pod \"kubeflow-trainer-controller-manager-55f5694779-lq9pl\" (UID: \"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8\") " pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.494733 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.494699 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:47.612491 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:47.612464 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl"] Apr 17 09:28:47.616524 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:28:47.616495 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62b1cf95_2ee8_450f_8cff_e5e7fb34a0e8.slice/crio-e47614d6b9ac64f201eca29b6892c9ebce0a52c7199a959264c0c45746ebe571 WatchSource:0}: Error finding container e47614d6b9ac64f201eca29b6892c9ebce0a52c7199a959264c0c45746ebe571: Status 404 returned error can't find the container with id e47614d6b9ac64f201eca29b6892c9ebce0a52c7199a959264c0c45746ebe571 Apr 17 09:28:48.084558 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:48.084525 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" event={"ID":"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8","Type":"ContainerStarted","Data":"e47614d6b9ac64f201eca29b6892c9ebce0a52c7199a959264c0c45746ebe571"} Apr 17 09:28:50.092307 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:50.092269 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" event={"ID":"62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8","Type":"ContainerStarted","Data":"062834d5f0f7431784f4eb2d3668eb4244cb1b1bb204b5fceb77a58641eb5e0b"} Apr 17 09:28:50.092752 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:50.092368 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:28:50.108918 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:28:50.108873 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" podStartSLOduration=0.759996458 podStartE2EDuration="3.108858782s" podCreationTimestamp="2026-04-17 09:28:47 +0000 UTC" firstStartedPulling="2026-04-17 09:28:47.618496808 +0000 UTC m=+477.431409962" lastFinishedPulling="2026-04-17 09:28:49.967359132 +0000 UTC m=+479.780272286" observedRunningTime="2026-04-17 09:28:50.108816277 +0000 UTC m=+479.921729449" watchObservedRunningTime="2026-04-17 09:28:50.108858782 +0000 UTC m=+479.921771957" Apr 17 09:29:06.100395 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:06.100361 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/kubeflow-trainer-controller-manager-55f5694779-lq9pl" Apr 17 09:29:06.972626 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:06.972587 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sssl6/must-gather-5vxvf"] Apr 17 09:29:06.975851 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:06.975835 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:06.978339 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:06.978313 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-sssl6\"/\"openshift-service-ca.crt\"" Apr 17 09:29:06.978437 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:06.978317 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-sssl6\"/\"kube-root-ca.crt\"" Apr 17 09:29:06.979330 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:06.979313 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-sssl6\"/\"default-dockercfg-8ppmx\"" Apr 17 09:29:06.986903 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:06.986881 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sssl6/must-gather-5vxvf"] Apr 17 09:29:07.119991 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.119954 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bvrs\" (UniqueName: \"kubernetes.io/projected/70255375-5649-4b9c-a6e0-0f5e3bbad69b-kube-api-access-6bvrs\") pod \"must-gather-5vxvf\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:07.120372 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.120041 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/70255375-5649-4b9c-a6e0-0f5e3bbad69b-must-gather-output\") pod \"must-gather-5vxvf\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:07.220433 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.220379 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/70255375-5649-4b9c-a6e0-0f5e3bbad69b-must-gather-output\") pod \"must-gather-5vxvf\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:07.220433 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.220436 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6bvrs\" (UniqueName: \"kubernetes.io/projected/70255375-5649-4b9c-a6e0-0f5e3bbad69b-kube-api-access-6bvrs\") pod \"must-gather-5vxvf\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:07.220735 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.220714 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/70255375-5649-4b9c-a6e0-0f5e3bbad69b-must-gather-output\") pod \"must-gather-5vxvf\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:07.228065 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.228012 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bvrs\" (UniqueName: \"kubernetes.io/projected/70255375-5649-4b9c-a6e0-0f5e3bbad69b-kube-api-access-6bvrs\") pod \"must-gather-5vxvf\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:07.285001 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.284965 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:29:07.400329 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:07.400299 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sssl6/must-gather-5vxvf"] Apr 17 09:29:07.403277 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:29:07.403247 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70255375_5649_4b9c_a6e0_0f5e3bbad69b.slice/crio-32296ddf732dfc435acd8551171a699219a072d1a90a36aca2cfeafb5cf219d0 WatchSource:0}: Error finding container 32296ddf732dfc435acd8551171a699219a072d1a90a36aca2cfeafb5cf219d0: Status 404 returned error can't find the container with id 32296ddf732dfc435acd8551171a699219a072d1a90a36aca2cfeafb5cf219d0 Apr 17 09:29:08.141649 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:08.141604 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sssl6/must-gather-5vxvf" event={"ID":"70255375-5649-4b9c-a6e0-0f5e3bbad69b","Type":"ContainerStarted","Data":"32296ddf732dfc435acd8551171a699219a072d1a90a36aca2cfeafb5cf219d0"} Apr 17 09:29:13.156569 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:13.156530 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sssl6/must-gather-5vxvf" event={"ID":"70255375-5649-4b9c-a6e0-0f5e3bbad69b","Type":"ContainerStarted","Data":"33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a"} Apr 17 09:29:13.156569 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:13.156572 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sssl6/must-gather-5vxvf" event={"ID":"70255375-5649-4b9c-a6e0-0f5e3bbad69b","Type":"ContainerStarted","Data":"c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3"} Apr 17 09:29:13.170758 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:13.170707 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sssl6/must-gather-5vxvf" podStartSLOduration=2.341730138 podStartE2EDuration="7.170692202s" podCreationTimestamp="2026-04-17 09:29:06 +0000 UTC" firstStartedPulling="2026-04-17 09:29:07.404888681 +0000 UTC m=+497.217801834" lastFinishedPulling="2026-04-17 09:29:12.233850746 +0000 UTC m=+502.046763898" observedRunningTime="2026-04-17 09:29:13.170480533 +0000 UTC m=+502.983393709" watchObservedRunningTime="2026-04-17 09:29:13.170692202 +0000 UTC m=+502.983605377" Apr 17 09:29:22.003846 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:22.003811 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_kubeflow-trainer-controller-manager-55f5694779-lq9pl_62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8/manager/0.log" Apr 17 09:29:22.453340 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:22.453314 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_kubeflow-trainer-controller-manager-55f5694779-lq9pl_62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8/manager/0.log" Apr 17 09:29:22.904688 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:22.904615 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_kubeflow-trainer-controller-manager-55f5694779-lq9pl_62b1cf95-2ee8-450f-8cff-e5e7fb34a0e8/manager/0.log" Apr 17 09:29:50.667671 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:50.667590 2581 scope.go:117] "RemoveContainer" containerID="02a8a3741ef07182d158b6d521d25495e78bb632912e0c5d5a8ef6c6e1a3754d" Apr 17 09:29:57.296275 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:57.296237 2581 generic.go:358] "Generic (PLEG): container finished" podID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerID="c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3" exitCode=0 Apr 17 09:29:57.296275 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:57.296278 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sssl6/must-gather-5vxvf" event={"ID":"70255375-5649-4b9c-a6e0-0f5e3bbad69b","Type":"ContainerDied","Data":"c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3"} Apr 17 09:29:57.296807 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:57.296590 2581 scope.go:117] "RemoveContainer" containerID="c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3" Apr 17 09:29:58.228216 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:29:58.228187 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sssl6_must-gather-5vxvf_70255375-5649-4b9c-a6e0-0f5e3bbad69b/gather/0.log" Apr 17 09:30:01.208991 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:01.208957 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-hbnq9_819cb802-41c1-426d-bcc4-c96157d9e5ed/global-pull-secret-syncer/0.log" Apr 17 09:30:01.381905 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:01.381879 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-gdhfx_31f8d7d4-b363-48ef-b10f-1ad81e1b496d/konnectivity-agent/0.log" Apr 17 09:30:01.404343 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:01.404313 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-129-70.ec2.internal_9af7c40fd2295f75411f633f4223f981/haproxy/0.log" Apr 17 09:30:03.556553 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.556514 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sssl6/must-gather-5vxvf"] Apr 17 09:30:03.557030 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.556724 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-must-gather-sssl6/must-gather-5vxvf" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerName="copy" containerID="cri-o://33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a" gracePeriod=2 Apr 17 09:30:03.558922 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.558895 2581 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sssl6/must-gather-5vxvf"] Apr 17 09:30:03.559078 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.559054 2581 status_manager.go:895] "Failed to get status for pod" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" pod="openshift-must-gather-sssl6/must-gather-5vxvf" err="pods \"must-gather-5vxvf\" is forbidden: User \"system:node:ip-10-0-129-70.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-sssl6\": no relationship found between node 'ip-10-0-129-70.ec2.internal' and this object" Apr 17 09:30:03.784932 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.784910 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sssl6_must-gather-5vxvf_70255375-5649-4b9c-a6e0-0f5e3bbad69b/copy/0.log" Apr 17 09:30:03.785251 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.785236 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:30:03.787239 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.787218 2581 status_manager.go:895] "Failed to get status for pod" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" pod="openshift-must-gather-sssl6/must-gather-5vxvf" err="pods \"must-gather-5vxvf\" is forbidden: User \"system:node:ip-10-0-129-70.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-sssl6\": no relationship found between node 'ip-10-0-129-70.ec2.internal' and this object" Apr 17 09:30:03.862612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.862535 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/70255375-5649-4b9c-a6e0-0f5e3bbad69b-must-gather-output\") pod \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " Apr 17 09:30:03.862743 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.862622 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bvrs\" (UniqueName: \"kubernetes.io/projected/70255375-5649-4b9c-a6e0-0f5e3bbad69b-kube-api-access-6bvrs\") pod \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\" (UID: \"70255375-5649-4b9c-a6e0-0f5e3bbad69b\") " Apr 17 09:30:03.864804 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.864752 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70255375-5649-4b9c-a6e0-0f5e3bbad69b-kube-api-access-6bvrs" (OuterVolumeSpecName: "kube-api-access-6bvrs") pod "70255375-5649-4b9c-a6e0-0f5e3bbad69b" (UID: "70255375-5649-4b9c-a6e0-0f5e3bbad69b"). InnerVolumeSpecName "kube-api-access-6bvrs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 09:30:03.864909 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.864877 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70255375-5649-4b9c-a6e0-0f5e3bbad69b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "70255375-5649-4b9c-a6e0-0f5e3bbad69b" (UID: "70255375-5649-4b9c-a6e0-0f5e3bbad69b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 09:30:03.964056 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.964024 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6bvrs\" (UniqueName: \"kubernetes.io/projected/70255375-5649-4b9c-a6e0-0f5e3bbad69b-kube-api-access-6bvrs\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:30:03.964056 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:03.964050 2581 reconciler_common.go:299] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/70255375-5649-4b9c-a6e0-0f5e3bbad69b-must-gather-output\") on node \"ip-10-0-129-70.ec2.internal\" DevicePath \"\"" Apr 17 09:30:04.314712 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.314674 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sssl6_must-gather-5vxvf_70255375-5649-4b9c-a6e0-0f5e3bbad69b/copy/0.log" Apr 17 09:30:04.314987 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.314963 2581 generic.go:358] "Generic (PLEG): container finished" podID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerID="33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a" exitCode=143 Apr 17 09:30:04.315045 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.315010 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sssl6/must-gather-5vxvf" Apr 17 09:30:04.315081 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.315035 2581 scope.go:117] "RemoveContainer" containerID="33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a" Apr 17 09:30:04.317133 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.317109 2581 status_manager.go:895] "Failed to get status for pod" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" pod="openshift-must-gather-sssl6/must-gather-5vxvf" err="pods \"must-gather-5vxvf\" is forbidden: User \"system:node:ip-10-0-129-70.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-sssl6\": no relationship found between node 'ip-10-0-129-70.ec2.internal' and this object" Apr 17 09:30:04.322379 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.322204 2581 scope.go:117] "RemoveContainer" containerID="c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3" Apr 17 09:30:04.324565 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.324542 2581 status_manager.go:895] "Failed to get status for pod" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" pod="openshift-must-gather-sssl6/must-gather-5vxvf" err="pods \"must-gather-5vxvf\" is forbidden: User \"system:node:ip-10-0-129-70.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-sssl6\": no relationship found between node 'ip-10-0-129-70.ec2.internal' and this object" Apr 17 09:30:04.334054 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.334035 2581 scope.go:117] "RemoveContainer" containerID="33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a" Apr 17 09:30:04.334328 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:30:04.334300 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a\": container with ID starting with 33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a not found: ID does not exist" containerID="33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a" Apr 17 09:30:04.334411 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.334342 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a"} err="failed to get container status \"33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a\": rpc error: code = NotFound desc = could not find container \"33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a\": container with ID starting with 33fb8376a399ce6b12055dde8ad80108e8a266193d3a56095df1166392ac195a not found: ID does not exist" Apr 17 09:30:04.334411 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.334367 2581 scope.go:117] "RemoveContainer" containerID="c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3" Apr 17 09:30:04.334617 ip-10-0-129-70 kubenswrapper[2581]: E0417 09:30:04.334601 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3\": container with ID starting with c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3 not found: ID does not exist" containerID="c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3" Apr 17 09:30:04.334653 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.334624 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3"} err="failed to get container status \"c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3\": rpc error: code = NotFound desc = could not find container \"c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3\": container with ID starting with c57041021a60778a6e01ab06e9181ea5737e20ab55e5d6f720e05ab666dd66a3 not found: ID does not exist" Apr 17 09:30:04.752991 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:04.752950 2581 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" path="/var/lib/kubelet/pods/70255375-5649-4b9c-a6e0-0f5e3bbad69b/volumes" Apr 17 09:30:05.079701 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:05.079622 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-h96rp_02ffc515-c207-4767-b17d-e7944d4310b3/node-exporter/0.log" Apr 17 09:30:05.098379 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:05.098328 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-h96rp_02ffc515-c207-4767-b17d-e7944d4310b3/kube-rbac-proxy/0.log" Apr 17 09:30:05.118740 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:05.118714 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-h96rp_02ffc515-c207-4767-b17d-e7944d4310b3/init-textfile/0.log" Apr 17 09:30:07.600243 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.600206 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg"] Apr 17 09:30:07.600612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.600427 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerName="copy" Apr 17 09:30:07.600612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.600437 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerName="copy" Apr 17 09:30:07.600612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.600451 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerName="gather" Apr 17 09:30:07.600612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.600456 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerName="gather" Apr 17 09:30:07.600612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.600499 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerName="copy" Apr 17 09:30:07.600612 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.600510 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="70255375-5649-4b9c-a6e0-0f5e3bbad69b" containerName="gather" Apr 17 09:30:07.605677 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.605657 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.608260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.608238 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-xkrgt\"/\"kube-root-ca.crt\"" Apr 17 09:30:07.608345 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.608270 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-xkrgt\"/\"openshift-service-ca.crt\"" Apr 17 09:30:07.609301 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.609286 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-xkrgt\"/\"default-dockercfg-v97xl\"" Apr 17 09:30:07.611879 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.611842 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg"] Apr 17 09:30:07.691350 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.691313 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nnj9\" (UniqueName: \"kubernetes.io/projected/701fe4f6-d5db-4e85-a9cb-045135a0d688-kube-api-access-2nnj9\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.691520 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.691369 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-proc\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.691520 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.691430 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-podres\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.691520 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.691466 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-sys\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.691520 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.691498 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-lib-modules\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.792789 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.792743 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-lib-modules\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.792979 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.792818 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2nnj9\" (UniqueName: \"kubernetes.io/projected/701fe4f6-d5db-4e85-a9cb-045135a0d688-kube-api-access-2nnj9\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.792979 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.792873 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-proc\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.792979 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.792898 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-podres\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.792979 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.792928 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-sys\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.792979 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.792942 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-lib-modules\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.792979 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.792960 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-proc\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.793260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.793002 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-sys\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.793260 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.793031 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/701fe4f6-d5db-4e85-a9cb-045135a0d688-podres\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.800279 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.800247 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nnj9\" (UniqueName: \"kubernetes.io/projected/701fe4f6-d5db-4e85-a9cb-045135a0d688-kube-api-access-2nnj9\") pod \"perf-node-gather-daemonset-vghzg\" (UID: \"701fe4f6-d5db-4e85-a9cb-045135a0d688\") " pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:07.916107 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:07.916010 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:08.031978 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.031944 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg"] Apr 17 09:30:08.034746 ip-10-0-129-70 kubenswrapper[2581]: W0417 09:30:08.034718 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod701fe4f6_d5db_4e85_a9cb_045135a0d688.slice/crio-648ee361e7cfe8a1ab127caf44152e4b912060fdb029601da1acc4dddce4055b WatchSource:0}: Error finding container 648ee361e7cfe8a1ab127caf44152e4b912060fdb029601da1acc4dddce4055b: Status 404 returned error can't find the container with id 648ee361e7cfe8a1ab127caf44152e4b912060fdb029601da1acc4dddce4055b Apr 17 09:30:08.300948 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.300919 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-bnzr4_c4061c0c-dd6a-4488-81f0-2b518afd6af0/dns/0.log" Apr 17 09:30:08.319691 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.319661 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-bnzr4_c4061c0c-dd6a-4488-81f0-2b518afd6af0/kube-rbac-proxy/0.log" Apr 17 09:30:08.327147 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.327120 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" event={"ID":"701fe4f6-d5db-4e85-a9cb-045135a0d688","Type":"ContainerStarted","Data":"02790b1d66c8019defcf8c92ffe60735a5ebdbae553b024c7a966f146bf88186"} Apr 17 09:30:08.327251 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.327153 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" event={"ID":"701fe4f6-d5db-4e85-a9cb-045135a0d688","Type":"ContainerStarted","Data":"648ee361e7cfe8a1ab127caf44152e4b912060fdb029601da1acc4dddce4055b"} Apr 17 09:30:08.327251 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.327177 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:08.342049 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.342003 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" podStartSLOduration=1.341990146 podStartE2EDuration="1.341990146s" podCreationTimestamp="2026-04-17 09:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 09:30:08.341613061 +0000 UTC m=+558.154526236" watchObservedRunningTime="2026-04-17 09:30:08.341990146 +0000 UTC m=+558.154903365" Apr 17 09:30:08.383344 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.383305 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-6h6cx_56322e26-7d25-4937-af43-be08d98d2866/dns-node-resolver/0.log" Apr 17 09:30:08.890415 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:08.890382 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-sj6zp_9b8cdaba-1bb6-47fd-b3c8-300ebe7fff31/node-ca/0.log" Apr 17 09:30:09.882457 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:09.882428 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-9mk5g_696f7ce6-641f-473d-b364-c01310172e36/serve-healthcheck-canary/0.log" Apr 17 09:30:10.297945 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:10.297922 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-jmq8j_e182d14d-d75c-42e5-851c-7797d94b6f39/kube-rbac-proxy/0.log" Apr 17 09:30:10.315197 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:10.315172 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-jmq8j_e182d14d-d75c-42e5-851c-7797d94b6f39/exporter/0.log" Apr 17 09:30:10.337641 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:10.337614 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-jmq8j_e182d14d-d75c-42e5-851c-7797d94b6f39/extractor/0.log" Apr 17 09:30:14.339370 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:14.339342 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-xkrgt/perf-node-gather-daemonset-vghzg" Apr 17 09:30:15.571013 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:15.570985 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-2h6z9_bd655abe-79ad-487e-adbe-b874b344be93/kube-multus-additional-cni-plugins/0.log" Apr 17 09:30:15.591255 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:15.591230 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-2h6z9_bd655abe-79ad-487e-adbe-b874b344be93/egress-router-binary-copy/0.log" Apr 17 09:30:15.609082 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:15.609060 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-2h6z9_bd655abe-79ad-487e-adbe-b874b344be93/cni-plugins/0.log" Apr 17 09:30:15.626544 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:15.626523 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-2h6z9_bd655abe-79ad-487e-adbe-b874b344be93/bond-cni-plugin/0.log" Apr 17 09:30:15.645222 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:15.645196 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-2h6z9_bd655abe-79ad-487e-adbe-b874b344be93/routeoverride-cni/0.log" Apr 17 09:30:15.665162 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:15.665139 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-2h6z9_bd655abe-79ad-487e-adbe-b874b344be93/whereabouts-cni-bincopy/0.log" Apr 17 09:30:15.685249 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:15.685225 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-2h6z9_bd655abe-79ad-487e-adbe-b874b344be93/whereabouts-cni/0.log" Apr 17 09:30:16.024139 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:16.024112 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s8tqz_6c744de0-b054-441d-a935-6063dac1784d/kube-multus/0.log" Apr 17 09:30:16.130684 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:16.130608 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-84m5n_504207a8-f3c9-45e3-a2cd-66458cf6eff8/network-metrics-daemon/0.log" Apr 17 09:30:16.150420 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:16.150397 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-84m5n_504207a8-f3c9-45e3-a2cd-66458cf6eff8/kube-rbac-proxy/0.log" Apr 17 09:30:16.916992 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:16.916965 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/ovn-controller/0.log" Apr 17 09:30:16.937547 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:16.937518 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/ovn-acl-logging/0.log" Apr 17 09:30:16.961189 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:16.961168 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/kube-rbac-proxy-node/0.log" Apr 17 09:30:16.984508 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:16.984483 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/kube-rbac-proxy-ovn-metrics/0.log" Apr 17 09:30:17.004866 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:17.004845 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/northd/0.log" Apr 17 09:30:17.029389 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:17.029359 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/nbdb/0.log" Apr 17 09:30:17.055139 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:17.055118 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/sbdb/0.log" Apr 17 09:30:17.159374 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:17.159280 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-64h7n_604b7c24-c2c4-45ab-9dbe-a714dd18cd14/ovnkube-controller/0.log" Apr 17 09:30:18.676930 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:18.676895 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-zmtbt_02351e99-a2a3-489f-b4dc-f8f12e779f0d/network-check-target-container/0.log" Apr 17 09:30:19.418500 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:19.418470 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-8r786_6160dcfd-b407-4815-bb22-a36a6d52799d/iptables-alerter/0.log" Apr 17 09:30:19.997416 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:19.997383 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-9t4gl_fd8d642b-833b-4048-a36c-25491ebf64f9/tuned/0.log" Apr 17 09:30:23.102440 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:23.102408 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-jqwtm_729c1d11-856e-4ad5-8a1c-46175de73a63/csi-driver/0.log" Apr 17 09:30:23.120042 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:23.120018 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-jqwtm_729c1d11-856e-4ad5-8a1c-46175de73a63/csi-node-driver-registrar/0.log" Apr 17 09:30:23.139114 ip-10-0-129-70 kubenswrapper[2581]: I0417 09:30:23.139089 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-jqwtm_729c1d11-856e-4ad5-8a1c-46175de73a63/csi-liveness-probe/0.log"